var/home/core/zuul-output/0000755000175000017500000000000015071424674014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015071437556015506 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004675271615071437545017732 0ustar rootrootOct 08 09:11:43 crc systemd[1]: Starting Kubernetes Kubelet... Oct 08 09:11:43 crc restorecon[4629]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:43 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 09:11:44 crc restorecon[4629]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 09:11:44 crc restorecon[4629]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 08 09:11:45 crc kubenswrapper[4744]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 09:11:45 crc kubenswrapper[4744]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 08 09:11:45 crc kubenswrapper[4744]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 09:11:45 crc kubenswrapper[4744]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 09:11:45 crc kubenswrapper[4744]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 08 09:11:45 crc kubenswrapper[4744]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.156707 4744 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163137 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163162 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163168 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163172 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163177 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163181 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163185 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163189 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163193 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163198 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163203 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163208 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163213 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163218 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163223 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163227 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163231 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163235 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163238 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163243 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163247 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163252 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163256 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163259 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163262 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163266 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163270 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163273 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163277 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163280 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163291 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163294 4744 feature_gate.go:330] unrecognized feature gate: Example Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163298 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163301 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163305 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163309 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163313 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163318 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163321 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163325 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163329 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163332 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163335 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163339 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163343 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163347 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163351 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163358 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163363 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163385 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163389 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163394 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163415 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163419 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163423 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163428 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163440 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163445 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163449 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163453 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163457 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163461 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163465 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163469 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163472 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163490 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163496 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163500 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163504 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163509 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.163513 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164412 4744 flags.go:64] FLAG: --address="0.0.0.0" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164430 4744 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164438 4744 flags.go:64] FLAG: --anonymous-auth="true" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164445 4744 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164455 4744 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164460 4744 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164468 4744 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164477 4744 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164485 4744 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164491 4744 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164497 4744 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164503 4744 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164508 4744 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164513 4744 flags.go:64] FLAG: --cgroup-root="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164518 4744 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164523 4744 flags.go:64] FLAG: --client-ca-file="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164529 4744 flags.go:64] FLAG: --cloud-config="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164534 4744 flags.go:64] FLAG: --cloud-provider="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164539 4744 flags.go:64] FLAG: --cluster-dns="[]" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164547 4744 flags.go:64] FLAG: --cluster-domain="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164552 4744 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164557 4744 flags.go:64] FLAG: --config-dir="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164563 4744 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164569 4744 flags.go:64] FLAG: --container-log-max-files="5" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164577 4744 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164582 4744 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164587 4744 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164592 4744 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164597 4744 flags.go:64] FLAG: --contention-profiling="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164601 4744 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164605 4744 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164611 4744 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164615 4744 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164620 4744 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164625 4744 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164628 4744 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164632 4744 flags.go:64] FLAG: --enable-load-reader="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164636 4744 flags.go:64] FLAG: --enable-server="true" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164641 4744 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164651 4744 flags.go:64] FLAG: --event-burst="100" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164657 4744 flags.go:64] FLAG: --event-qps="50" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164663 4744 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164668 4744 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164675 4744 flags.go:64] FLAG: --eviction-hard="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164682 4744 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164688 4744 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164693 4744 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164698 4744 flags.go:64] FLAG: --eviction-soft="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164704 4744 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164709 4744 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164714 4744 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164720 4744 flags.go:64] FLAG: --experimental-mounter-path="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164725 4744 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164731 4744 flags.go:64] FLAG: --fail-swap-on="true" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164737 4744 flags.go:64] FLAG: --feature-gates="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164744 4744 flags.go:64] FLAG: --file-check-frequency="20s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164749 4744 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164753 4744 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164758 4744 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164762 4744 flags.go:64] FLAG: --healthz-port="10248" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164766 4744 flags.go:64] FLAG: --help="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164771 4744 flags.go:64] FLAG: --hostname-override="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164775 4744 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164779 4744 flags.go:64] FLAG: --http-check-frequency="20s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164784 4744 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164788 4744 flags.go:64] FLAG: --image-credential-provider-config="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164792 4744 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164795 4744 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164799 4744 flags.go:64] FLAG: --image-service-endpoint="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164804 4744 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164808 4744 flags.go:64] FLAG: --kube-api-burst="100" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164812 4744 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164817 4744 flags.go:64] FLAG: --kube-api-qps="50" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164821 4744 flags.go:64] FLAG: --kube-reserved="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164826 4744 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164830 4744 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164834 4744 flags.go:64] FLAG: --kubelet-cgroups="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164838 4744 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164842 4744 flags.go:64] FLAG: --lock-file="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164847 4744 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164852 4744 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164856 4744 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164862 4744 flags.go:64] FLAG: --log-json-split-stream="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164886 4744 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164891 4744 flags.go:64] FLAG: --log-text-split-stream="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164896 4744 flags.go:64] FLAG: --logging-format="text" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164900 4744 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164905 4744 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164910 4744 flags.go:64] FLAG: --manifest-url="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164914 4744 flags.go:64] FLAG: --manifest-url-header="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164920 4744 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164924 4744 flags.go:64] FLAG: --max-open-files="1000000" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164929 4744 flags.go:64] FLAG: --max-pods="110" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164934 4744 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164938 4744 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164942 4744 flags.go:64] FLAG: --memory-manager-policy="None" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164946 4744 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164950 4744 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164954 4744 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164958 4744 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164970 4744 flags.go:64] FLAG: --node-status-max-images="50" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164974 4744 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164979 4744 flags.go:64] FLAG: --oom-score-adj="-999" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164985 4744 flags.go:64] FLAG: --pod-cidr="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.164993 4744 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165003 4744 flags.go:64] FLAG: --pod-manifest-path="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165008 4744 flags.go:64] FLAG: --pod-max-pids="-1" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165013 4744 flags.go:64] FLAG: --pods-per-core="0" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165017 4744 flags.go:64] FLAG: --port="10250" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165022 4744 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165026 4744 flags.go:64] FLAG: --provider-id="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165030 4744 flags.go:64] FLAG: --qos-reserved="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165036 4744 flags.go:64] FLAG: --read-only-port="10255" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165041 4744 flags.go:64] FLAG: --register-node="true" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165045 4744 flags.go:64] FLAG: --register-schedulable="true" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165059 4744 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165069 4744 flags.go:64] FLAG: --registry-burst="10" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165073 4744 flags.go:64] FLAG: --registry-qps="5" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165078 4744 flags.go:64] FLAG: --reserved-cpus="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165082 4744 flags.go:64] FLAG: --reserved-memory="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165088 4744 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165092 4744 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165097 4744 flags.go:64] FLAG: --rotate-certificates="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165101 4744 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165105 4744 flags.go:64] FLAG: --runonce="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165109 4744 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165113 4744 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165118 4744 flags.go:64] FLAG: --seccomp-default="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165122 4744 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165126 4744 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165130 4744 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165135 4744 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165139 4744 flags.go:64] FLAG: --storage-driver-password="root" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165144 4744 flags.go:64] FLAG: --storage-driver-secure="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165148 4744 flags.go:64] FLAG: --storage-driver-table="stats" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165152 4744 flags.go:64] FLAG: --storage-driver-user="root" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165157 4744 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165162 4744 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165166 4744 flags.go:64] FLAG: --system-cgroups="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165170 4744 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165177 4744 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165182 4744 flags.go:64] FLAG: --tls-cert-file="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165186 4744 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165192 4744 flags.go:64] FLAG: --tls-min-version="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165196 4744 flags.go:64] FLAG: --tls-private-key-file="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165200 4744 flags.go:64] FLAG: --topology-manager-policy="none" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165205 4744 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165209 4744 flags.go:64] FLAG: --topology-manager-scope="container" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165213 4744 flags.go:64] FLAG: --v="2" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165220 4744 flags.go:64] FLAG: --version="false" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165225 4744 flags.go:64] FLAG: --vmodule="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165231 4744 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165235 4744 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165361 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165387 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165391 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165396 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165400 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165403 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165407 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165410 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165414 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165419 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165423 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165427 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165431 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165435 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165439 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165442 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165447 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165451 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165454 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165458 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165461 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165465 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165468 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165471 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165475 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165478 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165482 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165485 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165489 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165492 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165495 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165501 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165505 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165508 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165513 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165517 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165522 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165527 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165532 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165537 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165541 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165546 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165550 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165556 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165560 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165574 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165579 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165583 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165589 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165593 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165597 4744 feature_gate.go:330] unrecognized feature gate: Example Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165601 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165606 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165610 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165614 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165618 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165622 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165627 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165631 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165636 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165640 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165644 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165649 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165653 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165656 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165660 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165663 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165667 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165670 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165673 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.165678 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.165691 4744 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.182548 4744 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.182620 4744 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182790 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182819 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182830 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182840 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182851 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182862 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182871 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182880 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182889 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182897 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182908 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182919 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182928 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182937 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182945 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182954 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182961 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182969 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182978 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182986 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.182994 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183001 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183009 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183018 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183025 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183032 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183041 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183051 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183060 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183069 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183079 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183088 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183096 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183104 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183115 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183123 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183131 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183140 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183148 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183157 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183164 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183175 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183187 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183196 4744 feature_gate.go:330] unrecognized feature gate: Example Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183205 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183213 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183221 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183229 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183236 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183245 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183253 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183260 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183268 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183276 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183284 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183292 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183303 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183313 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183323 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183332 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183342 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183351 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183360 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183391 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183400 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183408 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183416 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183424 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183431 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183439 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183450 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.183465 4744 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183713 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183729 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183740 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183751 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183761 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183771 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183781 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183792 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183802 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183815 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183824 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183834 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183842 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183850 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183858 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183866 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183874 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183885 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183894 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183902 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183910 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183918 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183926 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183933 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183941 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183949 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183956 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183964 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183971 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183979 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183987 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.183994 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184002 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184009 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184057 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184067 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184077 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184086 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184097 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184108 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184117 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184125 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184133 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184141 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184149 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184157 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184166 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184176 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184188 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184197 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184205 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184213 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184221 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184229 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184238 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184246 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184253 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184261 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184268 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184279 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184289 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184297 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184307 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184317 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184327 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184335 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184344 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184352 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184360 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184389 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.184399 4744 feature_gate.go:330] unrecognized feature gate: Example Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.184413 4744 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.186319 4744 server.go:940] "Client rotation is on, will bootstrap in background" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.192800 4744 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.192997 4744 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.195102 4744 server.go:997] "Starting client certificate rotation" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.195159 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.196997 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-16 20:43:27.180421982 +0000 UTC Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.197550 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1667h31m41.982881699s for next certificate rotation Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.223945 4744 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.227704 4744 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.248681 4744 log.go:25] "Validated CRI v1 runtime API" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.301082 4744 log.go:25] "Validated CRI v1 image API" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.303730 4744 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.312139 4744 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-08-09-06-07-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.312200 4744 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.336759 4744 manager.go:217] Machine: {Timestamp:2025-10-08 09:11:45.3329486 +0000 UTC m=+0.580593929 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2799998 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:f5297ca0-7781-4d45-97fe-e5122f26cce4 BootID:5c8c6e8b-9d7f-471f-a44b-640abd8b4177 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:5c:df:17 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:5c:df:17 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:38:95:a0 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:62:58:3a Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:40:cc:51 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:48:c9:7c Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:ad:f0:ba Speed:-1 Mtu:1496} {Name:eth10 MacAddress:2e:cb:fd:91:05:98 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ae:70:00:69:e0:c7 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.337210 4744 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.337553 4744 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.338199 4744 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.338590 4744 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.339002 4744 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.339333 4744 topology_manager.go:138] "Creating topology manager with none policy" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.339355 4744 container_manager_linux.go:303] "Creating device plugin manager" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.340306 4744 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.340352 4744 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.341427 4744 state_mem.go:36] "Initialized new in-memory state store" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.342028 4744 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.346337 4744 kubelet.go:418] "Attempting to sync node with API server" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.346395 4744 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.346437 4744 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.346458 4744 kubelet.go:324] "Adding apiserver pod source" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.346476 4744 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.354080 4744 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.355967 4744 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.357051 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.357264 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.357858 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.358157 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.359171 4744 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.361211 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.361263 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.361280 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.361298 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.361328 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.361343 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.361358 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.361412 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.361430 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.361446 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.361511 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.361529 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.362695 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.363615 4744 server.go:1280] "Started kubelet" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.365920 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.366411 4744 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.366432 4744 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 08 09:11:45 crc systemd[1]: Started Kubernetes Kubelet. Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.371426 4744 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.371729 4744 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.371940 4744 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 11:00:18.447036717 +0000 UTC Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.372042 4744 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1921h48m33.075002092s for next certificate rotation Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.372468 4744 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.374706 4744 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.374732 4744 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.375766 4744 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.376704 4744 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.376205 4744 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.217:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c7919a02e726a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-08 09:11:45.363563114 +0000 UTC m=+0.611208393,LastTimestamp:2025-10-08 09:11:45.363563114 +0000 UTC m=+0.611208393,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.378725 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="200ms" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.383851 4744 factory.go:55] Registering systemd factory Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.385044 4744 factory.go:221] Registration of the systemd container factory successfully Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.388839 4744 server.go:460] "Adding debug handlers to kubelet server" Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.388868 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.389045 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.390404 4744 factory.go:153] Registering CRI-O factory Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.390475 4744 factory.go:221] Registration of the crio container factory successfully Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.390650 4744 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.390712 4744 factory.go:103] Registering Raw factory Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.390747 4744 manager.go:1196] Started watching for new ooms in manager Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.391998 4744 manager.go:319] Starting recovery of all containers Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.397823 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.397876 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.397897 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.397911 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.397926 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.397940 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.397954 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.397970 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.397991 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.398008 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.398030 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.398045 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399675 4744 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399711 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399738 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399757 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399775 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399789 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399803 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399815 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399829 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399846 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399862 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399876 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399890 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399905 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399925 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399945 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.399962 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400029 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400045 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400059 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400072 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400086 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400100 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400113 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400128 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400142 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400156 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400169 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400183 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400197 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400212 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400231 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400248 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400266 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400285 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400302 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400324 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400341 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400357 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400401 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400458 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400487 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400507 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400526 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400546 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400564 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400581 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400599 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400619 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400637 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400659 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400676 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400694 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400715 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400735 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400755 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400778 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400796 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400814 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.400832 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401088 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401109 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401129 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401147 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401161 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401176 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401193 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401207 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401223 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401240 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401254 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401271 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401286 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401300 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401315 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401332 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401348 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401363 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401403 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401421 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401436 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401452 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401468 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401484 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401500 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401518 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401532 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401548 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401564 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401579 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401595 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401609 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401630 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401653 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401673 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401692 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401713 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401730 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401749 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401766 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401782 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401798 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401816 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401834 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401849 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401865 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401881 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401896 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401910 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401925 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401942 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401956 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401970 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.401986 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402005 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402021 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402040 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402055 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402071 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402087 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402103 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402121 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402136 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402154 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402176 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402199 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402215 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402234 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402252 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402270 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402288 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402306 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402327 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402344 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402441 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402463 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402482 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402501 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402519 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402537 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402553 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402570 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402585 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402602 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402620 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402636 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402650 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402663 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402674 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402686 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402699 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402714 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402727 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402747 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402765 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402793 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402817 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402839 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402855 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402874 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402890 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402906 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402919 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402932 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402945 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402957 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402970 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402985 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.402999 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403012 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403026 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403038 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403050 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403062 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403074 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403089 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403101 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403113 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403126 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403138 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403151 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403162 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403175 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403187 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403200 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403212 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403228 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403240 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403255 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403273 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403286 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403298 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403310 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403323 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403337 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403350 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403362 4744 reconstruct.go:97] "Volume reconstruction finished" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.403395 4744 reconciler.go:26] "Reconciler: start to sync state" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.420310 4744 manager.go:324] Recovery completed Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.438167 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.441520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.441581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.441595 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.442745 4744 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.442845 4744 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.442932 4744 state_mem.go:36] "Initialized new in-memory state store" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.446745 4744 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.451056 4744 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.451131 4744 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.451481 4744 kubelet.go:2335] "Starting kubelet main sync loop" Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.451910 4744 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.454061 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.454175 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.457206 4744 policy_none.go:49] "None policy: Start" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.458417 4744 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.458458 4744 state_mem.go:35] "Initializing new in-memory state store" Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.476852 4744 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.518812 4744 manager.go:334] "Starting Device Plugin manager" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.518888 4744 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.518906 4744 server.go:79] "Starting device plugin registration server" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.519577 4744 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.519595 4744 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.519793 4744 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.519984 4744 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.520001 4744 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.529794 4744 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.552825 4744 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.553305 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.555030 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.555186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.555348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.555669 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.555820 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.555890 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.557141 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.557206 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.557220 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.557468 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.557649 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.557715 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.557990 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.558030 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.558044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.558826 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.558863 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.558874 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.559010 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.559172 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.559216 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.559448 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.559576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.559653 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.560500 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.560571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.560595 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.560876 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.561001 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.561061 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.562235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.562263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.562276 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.562528 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.562548 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.562559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.562731 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.562770 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.562730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.563203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.563216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.564698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.564761 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.564784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.580303 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="400ms" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.608441 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.608523 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.608588 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.608818 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.608879 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.608907 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.608934 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.608962 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.608990 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.609069 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.609168 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.609237 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.609322 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.609407 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.609436 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.619888 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.621945 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.621991 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.622004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.622039 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.622638 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710338 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710530 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710591 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710635 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710688 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710722 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710768 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710786 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710690 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710741 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710639 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710888 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710733 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.710948 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711001 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711076 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711040 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711144 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711188 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711232 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711243 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711281 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711319 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711331 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711358 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711419 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711429 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711515 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.711601 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.822970 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.824672 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.824733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.824751 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.824794 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.825586 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.884468 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.892112 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.908166 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.916737 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: I1008 09:11:45.921237 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.939942 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-97d43b0cb4953d483de7176796a4040d30c069e7bf960f5a73978d307b0b8739 WatchSource:0}: Error finding container 97d43b0cb4953d483de7176796a4040d30c069e7bf960f5a73978d307b0b8739: Status 404 returned error can't find the container with id 97d43b0cb4953d483de7176796a4040d30c069e7bf960f5a73978d307b0b8739 Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.963009 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-f2a6ca85ff3a716d87d60b5fbd06f923e5b7551b0be0451bfe51658476d61a59 WatchSource:0}: Error finding container f2a6ca85ff3a716d87d60b5fbd06f923e5b7551b0be0451bfe51658476d61a59: Status 404 returned error can't find the container with id f2a6ca85ff3a716d87d60b5fbd06f923e5b7551b0be0451bfe51658476d61a59 Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.969733 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-4c750f029dd3952467db36a81dd2821c3cecf59e819f60513ac7fbf57dc30a05 WatchSource:0}: Error finding container 4c750f029dd3952467db36a81dd2821c3cecf59e819f60513ac7fbf57dc30a05: Status 404 returned error can't find the container with id 4c750f029dd3952467db36a81dd2821c3cecf59e819f60513ac7fbf57dc30a05 Oct 08 09:11:45 crc kubenswrapper[4744]: W1008 09:11:45.971325 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-1750a229acdf9054876a2c76a8d7c8caa5264ed5d062f6c5944f73ad3d18310f WatchSource:0}: Error finding container 1750a229acdf9054876a2c76a8d7c8caa5264ed5d062f6c5944f73ad3d18310f: Status 404 returned error can't find the container with id 1750a229acdf9054876a2c76a8d7c8caa5264ed5d062f6c5944f73ad3d18310f Oct 08 09:11:45 crc kubenswrapper[4744]: E1008 09:11:45.981555 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="800ms" Oct 08 09:11:46 crc kubenswrapper[4744]: I1008 09:11:46.226420 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:46 crc kubenswrapper[4744]: I1008 09:11:46.228936 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:46 crc kubenswrapper[4744]: I1008 09:11:46.229004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:46 crc kubenswrapper[4744]: I1008 09:11:46.229025 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:46 crc kubenswrapper[4744]: I1008 09:11:46.229067 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 09:11:46 crc kubenswrapper[4744]: E1008 09:11:46.229880 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Oct 08 09:11:46 crc kubenswrapper[4744]: I1008 09:11:46.366884 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:46 crc kubenswrapper[4744]: I1008 09:11:46.456801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f2a6ca85ff3a716d87d60b5fbd06f923e5b7551b0be0451bfe51658476d61a59"} Oct 08 09:11:46 crc kubenswrapper[4744]: I1008 09:11:46.459126 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1707af84a90dacfd86599f8a85745b68226958937bf73fbfbc10161be3fcbf5b"} Oct 08 09:11:46 crc kubenswrapper[4744]: I1008 09:11:46.460715 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"97d43b0cb4953d483de7176796a4040d30c069e7bf960f5a73978d307b0b8739"} Oct 08 09:11:46 crc kubenswrapper[4744]: I1008 09:11:46.462745 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1750a229acdf9054876a2c76a8d7c8caa5264ed5d062f6c5944f73ad3d18310f"} Oct 08 09:11:46 crc kubenswrapper[4744]: I1008 09:11:46.464720 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4c750f029dd3952467db36a81dd2821c3cecf59e819f60513ac7fbf57dc30a05"} Oct 08 09:11:46 crc kubenswrapper[4744]: E1008 09:11:46.783051 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="1.6s" Oct 08 09:11:46 crc kubenswrapper[4744]: W1008 09:11:46.862635 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:46 crc kubenswrapper[4744]: E1008 09:11:46.862753 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 08 09:11:46 crc kubenswrapper[4744]: W1008 09:11:46.894460 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:46 crc kubenswrapper[4744]: E1008 09:11:46.894620 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 08 09:11:46 crc kubenswrapper[4744]: W1008 09:11:46.951997 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:46 crc kubenswrapper[4744]: E1008 09:11:46.952195 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.030356 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:47 crc kubenswrapper[4744]: W1008 09:11:47.031348 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:47 crc kubenswrapper[4744]: E1008 09:11:47.031585 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.033521 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.033627 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.033656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.033719 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 09:11:47 crc kubenswrapper[4744]: E1008 09:11:47.034427 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.367810 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.483083 4744 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="d88f8a327e3a438414dacb96d94140f31f9218c7fac2b48e97834f61b229328f" exitCode=0 Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.483216 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"d88f8a327e3a438414dacb96d94140f31f9218c7fac2b48e97834f61b229328f"} Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.483243 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.484612 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.484659 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.484676 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.486408 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340" exitCode=0 Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.486563 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340"} Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.486593 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.487709 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.487751 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.487765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.489563 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.490289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.490319 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.490328 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.491135 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7"} Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.491174 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654"} Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.491188 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c"} Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.491202 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71"} Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.491191 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.492277 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.492301 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.492310 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.494283 4744 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4" exitCode=0 Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.494356 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4"} Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.494411 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.496107 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.496133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.496143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.497339 4744 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="aa4c3ccaa8b6863b9b478be4b68234a19167552eec7ed4a0e925d5c41364635d" exitCode=0 Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.497383 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"aa4c3ccaa8b6863b9b478be4b68234a19167552eec7ed4a0e925d5c41364635d"} Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.497519 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.500283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.500346 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:47 crc kubenswrapper[4744]: I1008 09:11:47.500356 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.366787 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:48 crc kubenswrapper[4744]: E1008 09:11:48.384680 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.217:6443: connect: connection refused" interval="3.2s" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.503811 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f"} Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.503860 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f"} Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.503876 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd"} Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.503904 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d"} Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.507090 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"49105d17503dca06928dea1ef7758199019bd0dea2af285149f0d81f06d6fba1"} Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.507120 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5e979533e4efeb3bf49255888b07c7aaf770c00a2fb105991b9ace6ce5422d5a"} Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.507133 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"64f870e3d341312ab4d6a6faad06746ae0fbac51d385a11767119039cb5440ac"} Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.507264 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.508784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.508824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.508838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.510519 4744 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c47a06f6760e9cb93e20d5d22de698ed55ed2ab0976e5603c81c82bff27e0bcf" exitCode=0 Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.510590 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c47a06f6760e9cb93e20d5d22de698ed55ed2ab0976e5603c81c82bff27e0bcf"} Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.510677 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.511963 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.511985 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.511996 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.513728 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.514244 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"bf594c3cd3cc11ac842c65691ebb0dfef51325813283e24905216a5848020dad"} Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.514290 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.514746 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.514784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.514804 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.515517 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.515556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.515572 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.635446 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.636884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.636926 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.636940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.636975 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 09:11:48 crc kubenswrapper[4744]: E1008 09:11:48.637563 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.217:6443: connect: connection refused" node="crc" Oct 08 09:11:48 crc kubenswrapper[4744]: I1008 09:11:48.793734 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:48 crc kubenswrapper[4744]: W1008 09:11:48.811334 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:48 crc kubenswrapper[4744]: E1008 09:11:48.811449 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 08 09:11:48 crc kubenswrapper[4744]: W1008 09:11:48.972199 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.217:6443: connect: connection refused Oct 08 09:11:48 crc kubenswrapper[4744]: E1008 09:11:48.972298 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.217:6443: connect: connection refused" logger="UnhandledError" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.131091 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.519753 4744 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="334cc52b46c353c9858318d8adc219ed21e6445b5d327fb740b6bb75f43c7585" exitCode=0 Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.519833 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"334cc52b46c353c9858318d8adc219ed21e6445b5d327fb740b6bb75f43c7585"} Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.519937 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.521180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.521222 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.521239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.525817 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.525939 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.525949 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.526219 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.525787 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972"} Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.527330 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.527405 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.527416 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.527590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.527625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.527638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.527798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.527821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.527833 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.528464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.528500 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:49 crc kubenswrapper[4744]: I1008 09:11:49.528512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.341229 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.532955 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.533420 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fd3cfb654417e4632d8e9e55a228e3fbbfd533336e0b12bc48e6a1d0e55899a7"} Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.533457 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"15ae2871346fa1eb7c7e9f2b82357e3c905ba1f71a4cee82e854df55b9a66cce"} Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.533468 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"64c803bf10d22bad88bf56c5daa9671ab060d4fce79053ec1467a9da9e95c32b"} Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.533476 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8847eaf636d48de2df9cb79f21b09c8a19b54c1b35a5efc38706982920574052"} Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.533545 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.533574 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.533890 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.533913 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.533923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.534738 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.534755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:50 crc kubenswrapper[4744]: I1008 09:11:50.534764 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.156235 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.296435 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.296639 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.298853 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.298906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.298920 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.541111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3058fd523955e5313b09d9da94639ea632a59777135663ffad664b0061331bc6"} Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.541139 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.541206 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.541240 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.542680 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.542741 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.542831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.542915 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.542976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.542999 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.838067 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.844146 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.844207 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.844227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:51 crc kubenswrapper[4744]: I1008 09:11:51.844264 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 09:11:52 crc kubenswrapper[4744]: I1008 09:11:52.544364 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:52 crc kubenswrapper[4744]: I1008 09:11:52.544411 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 09:11:52 crc kubenswrapper[4744]: I1008 09:11:52.544633 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:52 crc kubenswrapper[4744]: I1008 09:11:52.546491 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:52 crc kubenswrapper[4744]: I1008 09:11:52.546554 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:52 crc kubenswrapper[4744]: I1008 09:11:52.546500 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:52 crc kubenswrapper[4744]: I1008 09:11:52.546618 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:52 crc kubenswrapper[4744]: I1008 09:11:52.546638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:52 crc kubenswrapper[4744]: I1008 09:11:52.546580 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:53 crc kubenswrapper[4744]: I1008 09:11:53.071957 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:11:53 crc kubenswrapper[4744]: I1008 09:11:53.547619 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:53 crc kubenswrapper[4744]: I1008 09:11:53.549401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:53 crc kubenswrapper[4744]: I1008 09:11:53.549458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:53 crc kubenswrapper[4744]: I1008 09:11:53.549472 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:54 crc kubenswrapper[4744]: I1008 09:11:54.297023 4744 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 09:11:54 crc kubenswrapper[4744]: I1008 09:11:54.297160 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 09:11:55 crc kubenswrapper[4744]: I1008 09:11:55.042099 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 08 09:11:55 crc kubenswrapper[4744]: I1008 09:11:55.042438 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:55 crc kubenswrapper[4744]: I1008 09:11:55.043997 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:55 crc kubenswrapper[4744]: I1008 09:11:55.044081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:55 crc kubenswrapper[4744]: I1008 09:11:55.044104 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:55 crc kubenswrapper[4744]: E1008 09:11:55.530214 4744 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 09:11:55 crc kubenswrapper[4744]: I1008 09:11:55.882276 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:55 crc kubenswrapper[4744]: I1008 09:11:55.882674 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:55 crc kubenswrapper[4744]: I1008 09:11:55.885046 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:55 crc kubenswrapper[4744]: I1008 09:11:55.885107 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:55 crc kubenswrapper[4744]: I1008 09:11:55.885122 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:55 crc kubenswrapper[4744]: I1008 09:11:55.889083 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:56 crc kubenswrapper[4744]: I1008 09:11:56.551786 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:56 crc kubenswrapper[4744]: I1008 09:11:56.560611 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:11:56 crc kubenswrapper[4744]: I1008 09:11:56.561833 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:56 crc kubenswrapper[4744]: I1008 09:11:56.563222 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:56 crc kubenswrapper[4744]: I1008 09:11:56.563307 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:56 crc kubenswrapper[4744]: I1008 09:11:56.563326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:57 crc kubenswrapper[4744]: I1008 09:11:57.564726 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:11:57 crc kubenswrapper[4744]: I1008 09:11:57.566781 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:11:57 crc kubenswrapper[4744]: I1008 09:11:57.566825 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:11:57 crc kubenswrapper[4744]: I1008 09:11:57.566836 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:11:59 crc kubenswrapper[4744]: W1008 09:11:59.298655 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 08 09:11:59 crc kubenswrapper[4744]: I1008 09:11:59.298781 4744 trace.go:236] Trace[639624546]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 09:11:49.296) (total time: 10001ms): Oct 08 09:11:59 crc kubenswrapper[4744]: Trace[639624546]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:11:59.298) Oct 08 09:11:59 crc kubenswrapper[4744]: Trace[639624546]: [10.001773806s] [10.001773806s] END Oct 08 09:11:59 crc kubenswrapper[4744]: E1008 09:11:59.298814 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 08 09:11:59 crc kubenswrapper[4744]: I1008 09:11:59.368650 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 08 09:11:59 crc kubenswrapper[4744]: I1008 09:11:59.647887 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 08 09:11:59 crc kubenswrapper[4744]: I1008 09:11:59.647971 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 08 09:11:59 crc kubenswrapper[4744]: I1008 09:11:59.656138 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 08 09:11:59 crc kubenswrapper[4744]: I1008 09:11:59.656216 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.349976 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]log ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]etcd ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/generic-apiserver-start-informers ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/priority-and-fairness-filter ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/start-apiextensions-informers ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/start-apiextensions-controllers ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/crd-informer-synced ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/start-system-namespaces-controller ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 08 09:12:00 crc kubenswrapper[4744]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 08 09:12:00 crc kubenswrapper[4744]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/bootstrap-controller ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/start-kube-aggregator-informers ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/apiservice-registration-controller ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/apiservice-discovery-controller ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]autoregister-completion ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/apiservice-openapi-controller ok Oct 08 09:12:00 crc kubenswrapper[4744]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 08 09:12:00 crc kubenswrapper[4744]: livez check failed Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.350072 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.436857 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.437310 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.439262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.439334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.439443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.509900 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.574644 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.576152 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.576187 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.576197 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:00 crc kubenswrapper[4744]: I1008 09:12:00.591145 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 08 09:12:01 crc kubenswrapper[4744]: I1008 09:12:01.577756 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:12:01 crc kubenswrapper[4744]: I1008 09:12:01.578919 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:01 crc kubenswrapper[4744]: I1008 09:12:01.578957 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:01 crc kubenswrapper[4744]: I1008 09:12:01.578971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:03 crc kubenswrapper[4744]: I1008 09:12:03.123594 4744 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 08 09:12:04 crc kubenswrapper[4744]: I1008 09:12:04.297400 4744 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 09:12:04 crc kubenswrapper[4744]: I1008 09:12:04.298045 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 09:12:04 crc kubenswrapper[4744]: E1008 09:12:04.624848 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 08 09:12:04 crc kubenswrapper[4744]: I1008 09:12:04.632958 4744 trace.go:236] Trace[367006166]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 09:11:49.949) (total time: 14683ms): Oct 08 09:12:04 crc kubenswrapper[4744]: Trace[367006166]: ---"Objects listed" error: 14683ms (09:12:04.632) Oct 08 09:12:04 crc kubenswrapper[4744]: Trace[367006166]: [14.683342088s] [14.683342088s] END Oct 08 09:12:04 crc kubenswrapper[4744]: I1008 09:12:04.633002 4744 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 08 09:12:04 crc kubenswrapper[4744]: I1008 09:12:04.634099 4744 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 08 09:12:04 crc kubenswrapper[4744]: I1008 09:12:04.634173 4744 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 08 09:12:04 crc kubenswrapper[4744]: E1008 09:12:04.634286 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 08 09:12:04 crc kubenswrapper[4744]: I1008 09:12:04.635185 4744 trace.go:236] Trace[354699090]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 09:11:54.149) (total time: 10485ms): Oct 08 09:12:04 crc kubenswrapper[4744]: Trace[354699090]: ---"Objects listed" error: 10485ms (09:12:04.634) Oct 08 09:12:04 crc kubenswrapper[4744]: Trace[354699090]: [10.485636506s] [10.485636506s] END Oct 08 09:12:04 crc kubenswrapper[4744]: I1008 09:12:04.635237 4744 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 08 09:12:04 crc kubenswrapper[4744]: I1008 09:12:04.665249 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46778->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 08 09:12:04 crc kubenswrapper[4744]: I1008 09:12:04.665329 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46778->192.168.126.11:17697: read: connection reset by peer" Oct 08 09:12:04 crc kubenswrapper[4744]: I1008 09:12:04.665264 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46784->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 08 09:12:04 crc kubenswrapper[4744]: I1008 09:12:04.665422 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46784->192.168.126.11:17697: read: connection reset by peer" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.346119 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.346920 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.346994 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.353253 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.358413 4744 apiserver.go:52] "Watching apiserver" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.361875 4744 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.362255 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-rgsf6","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.362862 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.362995 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.362932 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.363199 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.363404 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.363581 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.363723 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.363767 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.363774 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.363871 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rgsf6" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.368142 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.368262 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.368758 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.369260 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.369311 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.369668 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.370363 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.370522 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.372457 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.372521 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.377229 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.377602 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.379203 4744 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.399989 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.403313 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.414455 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.426505 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.435250 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437427 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437470 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437497 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437522 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437542 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437565 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437584 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437606 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437633 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437652 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437669 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437686 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437703 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437721 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437738 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437758 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437776 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437797 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437816 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437836 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437857 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437877 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437895 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437877 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.437912 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438005 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438027 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438047 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438044 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438068 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438098 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438123 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438143 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438229 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438256 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438279 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438302 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438323 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438354 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438403 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438469 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438497 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438525 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438545 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438568 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438592 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438644 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438662 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438681 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438695 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438714 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438728 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438745 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438764 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438783 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438824 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438840 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438857 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438874 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438890 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438909 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438947 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438963 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438981 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438997 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439017 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439038 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439057 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439100 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439116 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439132 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439151 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439168 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439185 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439207 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439226 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439245 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439265 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439282 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439300 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439326 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439357 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439405 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439432 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439493 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439513 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439532 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439549 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439571 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439613 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439631 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439649 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439666 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439684 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439703 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439719 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439736 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439754 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439796 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439813 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439834 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439853 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439872 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439890 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439909 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439928 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439945 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439963 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439981 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440001 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440022 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440046 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440064 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440082 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440100 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440118 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440139 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440157 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440174 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440191 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440209 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440227 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440248 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440266 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440281 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440300 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440317 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440344 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440381 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440400 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440417 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440434 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440454 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440476 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440494 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440515 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440535 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440554 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440574 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440592 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440610 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440627 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440653 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440670 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440689 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440711 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440730 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440748 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440768 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440786 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440802 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440825 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440853 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440877 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440894 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440911 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440932 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440952 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440970 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440989 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441008 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441029 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441046 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441065 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441084 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441101 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441118 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441136 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441155 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441175 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441198 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441220 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441239 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441259 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441277 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441297 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441314 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441337 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441363 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441533 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441552 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441569 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441588 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441607 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441626 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441643 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441661 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441680 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441700 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441717 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441737 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441756 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441773 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441862 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441885 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441902 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441921 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441938 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441981 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442011 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442031 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442054 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442075 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442096 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442118 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxjrj\" (UniqueName: \"kubernetes.io/projected/305091b5-c89f-49b3-a180-4b304261fc7c-kube-api-access-pxjrj\") pod \"node-resolver-rgsf6\" (UID: \"305091b5-c89f-49b3-a180-4b304261fc7c\") " pod="openshift-dns/node-resolver-rgsf6" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442137 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442160 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442181 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442203 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442224 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442246 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442267 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/305091b5-c89f-49b3-a180-4b304261fc7c-hosts-file\") pod \"node-resolver-rgsf6\" (UID: \"305091b5-c89f-49b3-a180-4b304261fc7c\") " pod="openshift-dns/node-resolver-rgsf6" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442286 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442306 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442410 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442423 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442435 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.444215 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.454724 4744 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438886 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.468163 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438929 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.438970 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439000 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439297 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439435 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439630 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439740 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439819 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439858 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.439973 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440065 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440175 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440216 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.468302 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440279 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440487 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440502 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440672 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440947 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441144 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.440355 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441204 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.441604 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.442241 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.442507 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.443806 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.444284 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.444456 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.444603 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.444843 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.444875 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.445193 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.445288 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.445408 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.445454 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.445573 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.447177 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.447249 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.447463 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.447741 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.447759 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.447899 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.448070 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.448339 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.448351 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.448475 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.448808 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.449184 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.449306 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.449674 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.449731 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.450270 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.450274 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.450458 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.451238 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.451318 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.451336 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.451550 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.451667 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.451653 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.451767 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.451861 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.451900 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.452045 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.452214 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.452245 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.452408 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.452642 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.452880 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.452885 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.452700 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.453058 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.453097 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.453191 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.453291 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.453517 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.453528 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.453537 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.453551 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.453574 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.453881 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.454130 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.454414 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.454419 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.454638 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.455098 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.455140 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.455148 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.455192 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.455805 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.455991 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.456081 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.456088 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:12:05.956067066 +0000 UTC m=+21.203712295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.456321 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.457002 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.457142 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.457256 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.457410 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.457273 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.457556 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.457897 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.457966 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.458465 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.458791 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.459246 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.459619 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.459926 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.460162 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.460554 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.460876 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.459714 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.461193 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.461403 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.461536 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.461684 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.461697 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.461930 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.461943 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.462231 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.462308 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.462327 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.462479 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.462523 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.462603 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.463129 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.463352 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.463399 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.463491 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.463557 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.463639 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.463876 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.464448 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.464610 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.464689 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.464801 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.465151 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.465632 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.465928 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.466206 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.466354 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.466569 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.466870 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.466926 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.467191 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.467644 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.468092 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.468105 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.468834 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.468902 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.469012 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.469424 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.469636 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.469941 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.470054 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.470327 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.470516 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.470732 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:05.970706808 +0000 UTC m=+21.218352047 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.470891 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.470933 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.471207 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.471224 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.471899 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.474585 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.474656 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.474851 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.475030 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.475389 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.475311 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.477542 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.477941 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.482541 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.482563 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.482664 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.485198 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.485266 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.485324 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.485471 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.485883 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.486031 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.486552 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.487795 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.488716 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.490878 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.491170 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.491268 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.491591 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.491767 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.491926 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:05.991904058 +0000 UTC m=+21.239549297 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.492061 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.492188 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.497880 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.502126 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.502879 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.498904 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.499645 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.500529 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.503445 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.501144 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.503959 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.507384 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.508802 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.509559 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.510506 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.512236 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.513844 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.514912 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.515294 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.514221 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.516410 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.516884 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.517621 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.519194 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.519871 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.523860 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.525749 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.525834 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.525893 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.526024 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:06.02600912 +0000 UTC m=+21.273654359 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.526147 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.527897 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.529804 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.529911 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.530660 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.530715 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.530900 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.530937 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.531021 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:06.030979569 +0000 UTC m=+21.278624808 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.533719 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.535222 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.537703 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.538508 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.539795 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.540431 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.541844 4744 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.542594 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.543085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/305091b5-c89f-49b3-a180-4b304261fc7c-hosts-file\") pod \"node-resolver-rgsf6\" (UID: \"305091b5-c89f-49b3-a180-4b304261fc7c\") " pod="openshift-dns/node-resolver-rgsf6" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.543618 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.543765 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.543864 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxjrj\" (UniqueName: \"kubernetes.io/projected/305091b5-c89f-49b3-a180-4b304261fc7c-kube-api-access-pxjrj\") pod \"node-resolver-rgsf6\" (UID: \"305091b5-c89f-49b3-a180-4b304261fc7c\") " pod="openshift-dns/node-resolver-rgsf6" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.543951 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.543584 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/305091b5-c89f-49b3-a180-4b304261fc7c-hosts-file\") pod \"node-resolver-rgsf6\" (UID: \"305091b5-c89f-49b3-a180-4b304261fc7c\") " pod="openshift-dns/node-resolver-rgsf6" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.544234 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.544523 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.544660 4744 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.544727 4744 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.544792 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.544860 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545506 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545735 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545765 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545778 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545794 4744 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545805 4744 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545817 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545830 4744 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545844 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545853 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545866 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545848 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545877 4744 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545970 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545980 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.545992 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546007 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546018 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546028 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546038 4744 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546050 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546060 4744 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546069 4744 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546078 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546090 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546100 4744 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546110 4744 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546122 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546176 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546186 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546196 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546207 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546216 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546225 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546234 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546246 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546257 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546266 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546277 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546286 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546295 4744 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546304 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546314 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546323 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546335 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546344 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546355 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546365 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546623 4744 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546634 4744 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546650 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546660 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546669 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546681 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546691 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546701 4744 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546710 4744 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546721 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546730 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546739 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546749 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546761 4744 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546770 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546780 4744 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546788 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546799 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546808 4744 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546817 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546828 4744 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546838 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546848 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546857 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546868 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546876 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546885 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546894 4744 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546905 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546915 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546925 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546937 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546947 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546956 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546964 4744 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546975 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546984 4744 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.546994 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547003 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547015 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547024 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547033 4744 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547042 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547054 4744 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547064 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547074 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547086 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547095 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547104 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547115 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547126 4744 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547135 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547144 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547152 4744 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547164 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547173 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547182 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547192 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547203 4744 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547211 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547220 4744 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547230 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547240 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547249 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547258 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547271 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547281 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547290 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547299 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547310 4744 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547318 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547326 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547336 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547344 4744 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547353 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547362 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547385 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547395 4744 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547404 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547413 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547424 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547433 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547442 4744 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547450 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547461 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547470 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547478 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547488 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547496 4744 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547506 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547515 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547528 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547536 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547545 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547554 4744 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547566 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547575 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547586 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547594 4744 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547605 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547614 4744 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547622 4744 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547653 4744 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547663 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547670 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547679 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547691 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547701 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547712 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547721 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547733 4744 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547742 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547751 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547762 4744 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547775 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547784 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547793 4744 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547805 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547813 4744 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547824 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547832 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547842 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547851 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547860 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547869 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547881 4744 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547889 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547898 4744 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547909 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547918 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547927 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547935 4744 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547946 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547954 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547963 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547972 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547986 4744 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.547996 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.548006 4744 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.549386 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.549965 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.550432 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.553304 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.554687 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.555252 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.556560 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.557549 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.561310 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.562681 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.563747 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.571328 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.572676 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.573198 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.578026 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxjrj\" (UniqueName: \"kubernetes.io/projected/305091b5-c89f-49b3-a180-4b304261fc7c-kube-api-access-pxjrj\") pod \"node-resolver-rgsf6\" (UID: \"305091b5-c89f-49b3-a180-4b304261fc7c\") " pod="openshift-dns/node-resolver-rgsf6" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.580384 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.581150 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.582300 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.582816 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.583325 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.589893 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.589939 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.590766 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.591981 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.592461 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.592527 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.599509 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972" exitCode=255 Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.599895 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972"} Oct 08 09:12:05 crc kubenswrapper[4744]: E1008 09:12:05.617747 4744 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.618112 4744 scope.go:117] "RemoveContainer" containerID="e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.623089 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.629971 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-ttxzq"] Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.632708 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.638069 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.638429 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.638561 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.638589 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.638728 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.648360 4744 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.667689 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.677546 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.689733 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 09:12:05 crc kubenswrapper[4744]: W1008 09:12:05.690154 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-bc49e1d05a1fe85d5e611a7c544662b170c940ecde40595afeb8a1a4318c40fd WatchSource:0}: Error finding container bc49e1d05a1fe85d5e611a7c544662b170c940ecde40595afeb8a1a4318c40fd: Status 404 returned error can't find the container with id bc49e1d05a1fe85d5e611a7c544662b170c940ecde40595afeb8a1a4318c40fd Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.697940 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.701499 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rgsf6" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.705018 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.718722 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.735332 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: W1008 09:12:05.737086 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod305091b5_c89f_49b3_a180_4b304261fc7c.slice/crio-090a362fd2de9e81680c79ab55f8a2427316a6fe78fb9243a3d92669b919e82a WatchSource:0}: Error finding container 090a362fd2de9e81680c79ab55f8a2427316a6fe78fb9243a3d92669b919e82a: Status 404 returned error can't find the container with id 090a362fd2de9e81680c79ab55f8a2427316a6fe78fb9243a3d92669b919e82a Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.749014 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7mf2\" (UniqueName: \"kubernetes.io/projected/9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb-kube-api-access-t7mf2\") pod \"machine-config-daemon-ttxzq\" (UID: \"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\") " pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.749086 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb-rootfs\") pod \"machine-config-daemon-ttxzq\" (UID: \"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\") " pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.749122 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb-mcd-auth-proxy-config\") pod \"machine-config-daemon-ttxzq\" (UID: \"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\") " pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.749169 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb-proxy-tls\") pod \"machine-config-daemon-ttxzq\" (UID: \"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\") " pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.750433 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: W1008 09:12:05.756892 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-c0f456a8b27957303d5bc02787a24a2d87a2c987afba313d8de6731765bd74eb WatchSource:0}: Error finding container c0f456a8b27957303d5bc02787a24a2d87a2c987afba313d8de6731765bd74eb: Status 404 returned error can't find the container with id c0f456a8b27957303d5bc02787a24a2d87a2c987afba313d8de6731765bd74eb Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.765998 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.779607 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.790795 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.805864 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.819211 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.836122 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.850205 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb-proxy-tls\") pod \"machine-config-daemon-ttxzq\" (UID: \"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\") " pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.850260 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb-rootfs\") pod \"machine-config-daemon-ttxzq\" (UID: \"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\") " pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.850283 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7mf2\" (UniqueName: \"kubernetes.io/projected/9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb-kube-api-access-t7mf2\") pod \"machine-config-daemon-ttxzq\" (UID: \"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\") " pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.850304 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb-mcd-auth-proxy-config\") pod \"machine-config-daemon-ttxzq\" (UID: \"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\") " pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.850798 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb-rootfs\") pod \"machine-config-daemon-ttxzq\" (UID: \"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\") " pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.851989 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb-mcd-auth-proxy-config\") pod \"machine-config-daemon-ttxzq\" (UID: \"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\") " pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.855747 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.863031 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb-proxy-tls\") pod \"machine-config-daemon-ttxzq\" (UID: \"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\") " pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.875969 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.877282 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7mf2\" (UniqueName: \"kubernetes.io/projected/9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb-kube-api-access-t7mf2\") pod \"machine-config-daemon-ttxzq\" (UID: \"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\") " pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: I1008 09:12:05.952713 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:12:05 crc kubenswrapper[4744]: W1008 09:12:05.980448 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d1b4cf3_79bd_4fc9_8718_60c1ea87fadb.slice/crio-986c077dc68497908096d986e8e106c8d577981a5f52e4e522380581b162f7ad WatchSource:0}: Error finding container 986c077dc68497908096d986e8e106c8d577981a5f52e4e522380581b162f7ad: Status 404 returned error can't find the container with id 986c077dc68497908096d986e8e106c8d577981a5f52e4e522380581b162f7ad Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.031474 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-dldlq"] Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.031815 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.032581 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-f6z7m"] Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.033250 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.034805 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.035214 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.035400 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.035621 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.052017 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.052268 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.052394 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.053983 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054151 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054204 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:07.054184625 +0000 UTC m=+22.301829864 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054334 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:12:07.054325919 +0000 UTC m=+22.301971158 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.054058 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.054558 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.054580 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.054600 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054689 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054707 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054717 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054744 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:07.054737152 +0000 UTC m=+22.302382391 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054792 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054801 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054808 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054832 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:07.054826854 +0000 UTC m=+22.302472083 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054878 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.054901 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:07.054893016 +0000 UTC m=+22.302538255 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.070676 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.088019 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.110758 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.125466 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.142495 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.155795 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-etc-kubernetes\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.155897 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-hostroot\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.155916 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bfffa6a9-a014-46b8-b4de-3261522cc21e-cnibin\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.155933 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bfffa6a9-a014-46b8-b4de-3261522cc21e-cni-binary-copy\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.155958 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7ea48507-1f33-46fc-ab34-c6e083dd86a9-cni-binary-copy\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.155981 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-multus-conf-dir\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156000 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-var-lib-kubelet\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156102 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bfffa6a9-a014-46b8-b4de-3261522cc21e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156251 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-run-k8s-cni-cncf-io\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156308 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bfffa6a9-a014-46b8-b4de-3261522cc21e-system-cni-dir\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156326 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bfffa6a9-a014-46b8-b4de-3261522cc21e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156372 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bfffa6a9-a014-46b8-b4de-3261522cc21e-os-release\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156409 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctjzs\" (UniqueName: \"kubernetes.io/projected/bfffa6a9-a014-46b8-b4de-3261522cc21e-kube-api-access-ctjzs\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156427 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7ea48507-1f33-46fc-ab34-c6e083dd86a9-multus-daemon-config\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156465 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n4nk\" (UniqueName: \"kubernetes.io/projected/7ea48507-1f33-46fc-ab34-c6e083dd86a9-kube-api-access-6n4nk\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156510 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-multus-cni-dir\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156534 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-cnibin\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156551 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-os-release\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156572 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-run-netns\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156592 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-var-lib-cni-bin\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156622 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-system-cni-dir\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156640 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-multus-socket-dir-parent\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156670 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-var-lib-cni-multus\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.156690 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-run-multus-certs\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.162189 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.172993 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.185181 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.197585 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.210538 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.221657 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.232300 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.243196 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.253773 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257034 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-etc-kubernetes\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257081 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-hostroot\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257109 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bfffa6a9-a014-46b8-b4de-3261522cc21e-cni-binary-copy\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257129 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bfffa6a9-a014-46b8-b4de-3261522cc21e-cnibin\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257146 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7ea48507-1f33-46fc-ab34-c6e083dd86a9-cni-binary-copy\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257163 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-multus-conf-dir\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257175 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bfffa6a9-a014-46b8-b4de-3261522cc21e-cnibin\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257179 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bfffa6a9-a014-46b8-b4de-3261522cc21e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257200 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-var-lib-kubelet\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257216 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bfffa6a9-a014-46b8-b4de-3261522cc21e-system-cni-dir\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257232 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bfffa6a9-a014-46b8-b4de-3261522cc21e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257247 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-run-k8s-cni-cncf-io\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257264 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bfffa6a9-a014-46b8-b4de-3261522cc21e-os-release\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257280 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctjzs\" (UniqueName: \"kubernetes.io/projected/bfffa6a9-a014-46b8-b4de-3261522cc21e-kube-api-access-ctjzs\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257296 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-multus-cni-dir\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257313 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-cnibin\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257351 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-os-release\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257369 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-run-netns\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257385 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7ea48507-1f33-46fc-ab34-c6e083dd86a9-multus-daemon-config\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257414 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n4nk\" (UniqueName: \"kubernetes.io/projected/7ea48507-1f33-46fc-ab34-c6e083dd86a9-kube-api-access-6n4nk\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257440 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-system-cni-dir\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257455 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-multus-socket-dir-parent\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257472 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-var-lib-cni-bin\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257502 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-run-multus-certs\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257535 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-var-lib-cni-multus\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257593 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-var-lib-cni-multus\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257128 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-etc-kubernetes\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257836 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bfffa6a9-a014-46b8-b4de-3261522cc21e-cni-binary-copy\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257910 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-multus-cni-dir\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257149 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-hostroot\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.257952 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-multus-conf-dir\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258128 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7ea48507-1f33-46fc-ab34-c6e083dd86a9-cni-binary-copy\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258177 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-cnibin\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258355 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bfffa6a9-a014-46b8-b4de-3261522cc21e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258417 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-var-lib-kubelet\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258450 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bfffa6a9-a014-46b8-b4de-3261522cc21e-system-cni-dir\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258466 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-os-release\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258494 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-run-netns\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258523 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-run-k8s-cni-cncf-io\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258649 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-system-cni-dir\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258712 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-multus-socket-dir-parent\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258747 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-var-lib-cni-bin\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258778 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7ea48507-1f33-46fc-ab34-c6e083dd86a9-host-run-multus-certs\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258507 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bfffa6a9-a014-46b8-b4de-3261522cc21e-os-release\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.258965 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7ea48507-1f33-46fc-ab34-c6e083dd86a9-multus-daemon-config\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.259357 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bfffa6a9-a014-46b8-b4de-3261522cc21e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.265322 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.274764 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.276933 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctjzs\" (UniqueName: \"kubernetes.io/projected/bfffa6a9-a014-46b8-b4de-3261522cc21e-kube-api-access-ctjzs\") pod \"multus-additional-cni-plugins-f6z7m\" (UID: \"bfffa6a9-a014-46b8-b4de-3261522cc21e\") " pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.277052 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n4nk\" (UniqueName: \"kubernetes.io/projected/7ea48507-1f33-46fc-ab34-c6e083dd86a9-kube-api-access-6n4nk\") pod \"multus-dldlq\" (UID: \"7ea48507-1f33-46fc-ab34-c6e083dd86a9\") " pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.284080 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.293124 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.303331 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.316815 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.329552 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.395488 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tmmnj"] Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.401767 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.407386 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.407582 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.407684 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.408224 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.408274 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.408531 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.409434 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.413696 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-dldlq" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.419555 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.423690 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: W1008 09:12:06.442964 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ea48507_1f33_46fc_ab34_c6e083dd86a9.slice/crio-d45dc537e3408d89b85a9b58ff7de8fef69e6d6869a82c3d064d351818ba84c2 WatchSource:0}: Error finding container d45dc537e3408d89b85a9b58ff7de8fef69e6d6869a82c3d064d351818ba84c2: Status 404 returned error can't find the container with id d45dc537e3408d89b85a9b58ff7de8fef69e6d6869a82c3d064d351818ba84c2 Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.443549 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.452729 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:06 crc kubenswrapper[4744]: E1008 09:12:06.452936 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.459284 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-systemd\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.459488 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-log-socket\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.459589 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-env-overrides\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.459687 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-run-ovn-kubernetes\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.459769 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-run-netns\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.459853 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovnkube-config\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.459923 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovn-node-metrics-cert\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.459999 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-systemd-units\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460086 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-kubelet\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460160 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-etc-openvswitch\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460224 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-cni-bin\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460294 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovnkube-script-lib\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460359 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-slash\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460440 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-openvswitch\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460428 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460529 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q28tm\" (UniqueName: \"kubernetes.io/projected/5a9bf215-f18b-4732-b12c-0355a3a0a15f-kube-api-access-q28tm\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460706 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-ovn\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460731 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-node-log\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460792 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-var-lib-openvswitch\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460817 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-cni-netd\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.460842 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.472585 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.484357 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.495103 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.509031 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.531704 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.551854 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.561820 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-node-log\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.561878 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-ovn\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.561912 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-var-lib-openvswitch\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.561937 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-cni-netd\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.561949 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-node-log\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.561962 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.561995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-systemd\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562010 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-var-lib-openvswitch\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562024 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-log-socket\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562042 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-ovn\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562048 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-env-overrides\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562073 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-run-ovn-kubernetes\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562096 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-run-netns\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562146 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovn-node-metrics-cert\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562168 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-systemd-units\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562191 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovnkube-config\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562213 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-kubelet\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562234 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-etc-openvswitch\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-cni-bin\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562275 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovnkube-script-lib\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562295 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-slash\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562318 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-openvswitch\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562340 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q28tm\" (UniqueName: \"kubernetes.io/projected/5a9bf215-f18b-4732-b12c-0355a3a0a15f-kube-api-access-q28tm\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562731 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-run-ovn-kubernetes\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562074 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562790 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-env-overrides\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562801 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-run-netns\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562099 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-cni-netd\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562890 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-systemd\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562925 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-log-socket\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562952 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-etc-openvswitch\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.562974 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-systemd-units\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.563480 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovnkube-config\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.563493 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-slash\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.563524 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-kubelet\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.563546 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-cni-bin\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.563602 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-openvswitch\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.563922 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovnkube-script-lib\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.567604 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovn-node-metrics-cert\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.575529 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.584738 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q28tm\" (UniqueName: \"kubernetes.io/projected/5a9bf215-f18b-4732-b12c-0355a3a0a15f-kube-api-access-q28tm\") pod \"ovnkube-node-tmmnj\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.589083 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.601746 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.608007 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.608072 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.608084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"986c077dc68497908096d986e8e106c8d577981a5f52e4e522380581b162f7ad"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.610121 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c0f456a8b27957303d5bc02787a24a2d87a2c987afba313d8de6731765bd74eb"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.614499 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.614555 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"bc49e1d05a1fe85d5e611a7c544662b170c940ecde40595afeb8a1a4318c40fd"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.619375 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dldlq" event={"ID":"7ea48507-1f33-46fc-ab34-c6e083dd86a9","Type":"ContainerStarted","Data":"d45dc537e3408d89b85a9b58ff7de8fef69e6d6869a82c3d064d351818ba84c2"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.625145 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.625210 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.625227 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c927aac1ec68e5d75cbb04794ab448a9e8e9bc0b33c7e146552a466b0db56436"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.628526 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rgsf6" event={"ID":"305091b5-c89f-49b3-a180-4b304261fc7c","Type":"ContainerStarted","Data":"84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.628598 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rgsf6" event={"ID":"305091b5-c89f-49b3-a180-4b304261fc7c","Type":"ContainerStarted","Data":"090a362fd2de9e81680c79ab55f8a2427316a6fe78fb9243a3d92669b919e82a"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.632835 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.635269 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.635620 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.637114 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" event={"ID":"bfffa6a9-a014-46b8-b4de-3261522cc21e","Type":"ContainerStarted","Data":"7b2b367f228baaba589682e7a7896b386a16312b523e800e819143085f8a3bb7"} Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.637811 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.677746 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.717358 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.722207 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 09:12:06 crc kubenswrapper[4744]: W1008 09:12:06.731762 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a9bf215_f18b_4732_b12c_0355a3a0a15f.slice/crio-bda60c6f4059e554ef5635a19438017eaf5b62aa6fadef4522cdcf59850de381 WatchSource:0}: Error finding container bda60c6f4059e554ef5635a19438017eaf5b62aa6fadef4522cdcf59850de381: Status 404 returned error can't find the container with id bda60c6f4059e554ef5635a19438017eaf5b62aa6fadef4522cdcf59850de381 Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.780227 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.823049 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.855075 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.892200 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.927701 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:06 crc kubenswrapper[4744]: I1008 09:12:06.963355 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.000291 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:06Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.037388 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.067895 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068001 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:12:09.0679779 +0000 UTC m=+24.315623139 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.068030 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.068085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.068113 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.068132 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068195 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068206 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068213 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068228 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068241 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:09.068234638 +0000 UTC m=+24.315879877 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068261 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:09.068251518 +0000 UTC m=+24.315896757 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068491 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068501 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068511 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068533 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:09.068526636 +0000 UTC m=+24.316171875 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068551 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.068577 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:09.068570108 +0000 UTC m=+24.316215347 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.079400 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.119790 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.161505 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.209784 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.248507 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.283591 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.325431 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.368067 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.402196 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.449290 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.452452 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.452546 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.452685 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:07 crc kubenswrapper[4744]: E1008 09:12:07.452769 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.457143 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.457871 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.458595 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.459523 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.460120 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.460724 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.461338 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.461980 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.462642 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.479278 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.527481 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.567371 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.642275 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dldlq" event={"ID":"7ea48507-1f33-46fc-ab34-c6e083dd86a9","Type":"ContainerStarted","Data":"706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f"} Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.643842 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321" exitCode=0 Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.643933 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321"} Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.643967 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"bda60c6f4059e554ef5635a19438017eaf5b62aa6fadef4522cdcf59850de381"} Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.645756 4744 generic.go:334] "Generic (PLEG): container finished" podID="bfffa6a9-a014-46b8-b4de-3261522cc21e" containerID="8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f" exitCode=0 Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.645878 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" event={"ID":"bfffa6a9-a014-46b8-b4de-3261522cc21e","Type":"ContainerDied","Data":"8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f"} Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.676896 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.700361 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.732161 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.747318 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.789222 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.822117 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.850523 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.894571 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.926823 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:07 crc kubenswrapper[4744]: I1008 09:12:07.961176 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:07Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.011777 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.042062 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.135257 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.166735 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.193577 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.216553 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.248878 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.287761 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.322860 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.359671 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.402518 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.419034 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-j2skv"] Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.419497 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-j2skv" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.445816 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.451675 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.452773 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:08 crc kubenswrapper[4744]: E1008 09:12:08.452993 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.469906 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.491035 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.511066 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.520072 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/56790493-3c4e-457d-bf41-aa9a7fb680ec-serviceca\") pod \"node-ca-j2skv\" (UID: \"56790493-3c4e-457d-bf41-aa9a7fb680ec\") " pod="openshift-image-registry/node-ca-j2skv" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.520350 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/56790493-3c4e-457d-bf41-aa9a7fb680ec-host\") pod \"node-ca-j2skv\" (UID: \"56790493-3c4e-457d-bf41-aa9a7fb680ec\") " pod="openshift-image-registry/node-ca-j2skv" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.520453 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l689j\" (UniqueName: \"kubernetes.io/projected/56790493-3c4e-457d-bf41-aa9a7fb680ec-kube-api-access-l689j\") pod \"node-ca-j2skv\" (UID: \"56790493-3c4e-457d-bf41-aa9a7fb680ec\") " pod="openshift-image-registry/node-ca-j2skv" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.562321 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.602540 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.621112 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/56790493-3c4e-457d-bf41-aa9a7fb680ec-host\") pod \"node-ca-j2skv\" (UID: \"56790493-3c4e-457d-bf41-aa9a7fb680ec\") " pod="openshift-image-registry/node-ca-j2skv" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.621258 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/56790493-3c4e-457d-bf41-aa9a7fb680ec-host\") pod \"node-ca-j2skv\" (UID: \"56790493-3c4e-457d-bf41-aa9a7fb680ec\") " pod="openshift-image-registry/node-ca-j2skv" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.621430 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l689j\" (UniqueName: \"kubernetes.io/projected/56790493-3c4e-457d-bf41-aa9a7fb680ec-kube-api-access-l689j\") pod \"node-ca-j2skv\" (UID: \"56790493-3c4e-457d-bf41-aa9a7fb680ec\") " pod="openshift-image-registry/node-ca-j2skv" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.621605 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/56790493-3c4e-457d-bf41-aa9a7fb680ec-serviceca\") pod \"node-ca-j2skv\" (UID: \"56790493-3c4e-457d-bf41-aa9a7fb680ec\") " pod="openshift-image-registry/node-ca-j2skv" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.623084 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/56790493-3c4e-457d-bf41-aa9a7fb680ec-serviceca\") pod \"node-ca-j2skv\" (UID: \"56790493-3c4e-457d-bf41-aa9a7fb680ec\") " pod="openshift-image-registry/node-ca-j2skv" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.641531 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.651453 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a"} Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.653619 4744 generic.go:334] "Generic (PLEG): container finished" podID="bfffa6a9-a014-46b8-b4de-3261522cc21e" containerID="f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b" exitCode=0 Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.653684 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" event={"ID":"bfffa6a9-a014-46b8-b4de-3261522cc21e","Type":"ContainerDied","Data":"f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b"} Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.660852 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e"} Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.661047 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad"} Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.661066 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72"} Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.661084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb"} Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.661094 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1"} Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.661104 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe"} Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.673284 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l689j\" (UniqueName: \"kubernetes.io/projected/56790493-3c4e-457d-bf41-aa9a7fb680ec-kube-api-access-l689j\") pod \"node-ca-j2skv\" (UID: \"56790493-3c4e-457d-bf41-aa9a7fb680ec\") " pod="openshift-image-registry/node-ca-j2skv" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.707503 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.735186 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-j2skv" Oct 08 09:12:08 crc kubenswrapper[4744]: W1008 09:12:08.751621 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56790493_3c4e_457d_bf41_aa9a7fb680ec.slice/crio-24251add9a77cbc90815b878362f9c63ec16fe3d82183295c95557c167de072e WatchSource:0}: Error finding container 24251add9a77cbc90815b878362f9c63ec16fe3d82183295c95557c167de072e: Status 404 returned error can't find the container with id 24251add9a77cbc90815b878362f9c63ec16fe3d82183295c95557c167de072e Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.754540 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.781759 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.819749 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.857561 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.897438 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.943732 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:08 crc kubenswrapper[4744]: I1008 09:12:08.980792 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:08Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.021729 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.058037 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.097363 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.127149 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.127254 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.127282 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.127306 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.127323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.127456 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.127512 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:13.127495474 +0000 UTC m=+28.375140713 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.127978 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:12:13.127970418 +0000 UTC m=+28.375615657 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.128036 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.128052 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.128065 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.128088 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:13.128082352 +0000 UTC m=+28.375727591 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.128123 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.128132 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.128139 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.128160 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:13.128150404 +0000 UTC m=+28.375795643 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.128205 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.128224 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:13.128218837 +0000 UTC m=+28.375864076 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.138396 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.183617 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.225106 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.260446 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.303836 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.338198 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.381263 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.424656 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.452594 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.452618 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.452857 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:09 crc kubenswrapper[4744]: E1008 09:12:09.453131 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.466419 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.501544 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.544453 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.594253 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.625794 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.666153 4744 generic.go:334] "Generic (PLEG): container finished" podID="bfffa6a9-a014-46b8-b4de-3261522cc21e" containerID="94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22" exitCode=0 Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.666279 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" event={"ID":"bfffa6a9-a014-46b8-b4de-3261522cc21e","Type":"ContainerDied","Data":"94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22"} Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.667873 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-j2skv" event={"ID":"56790493-3c4e-457d-bf41-aa9a7fb680ec","Type":"ContainerStarted","Data":"5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a"} Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.667948 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-j2skv" event={"ID":"56790493-3c4e-457d-bf41-aa9a7fb680ec","Type":"ContainerStarted","Data":"24251add9a77cbc90815b878362f9c63ec16fe3d82183295c95557c167de072e"} Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.670710 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.706144 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.739701 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.782172 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.823083 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.860632 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.898083 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.939362 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:09 crc kubenswrapper[4744]: I1008 09:12:09.981071 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:09Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.023781 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.061235 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.100661 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.138773 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.178228 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.452934 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:10 crc kubenswrapper[4744]: E1008 09:12:10.453109 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.677713 4744 generic.go:334] "Generic (PLEG): container finished" podID="bfffa6a9-a014-46b8-b4de-3261522cc21e" containerID="d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99" exitCode=0 Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.677859 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" event={"ID":"bfffa6a9-a014-46b8-b4de-3261522cc21e","Type":"ContainerDied","Data":"d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99"} Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.704635 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.728954 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.746266 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.759218 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.773487 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.784166 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.798623 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.816197 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.835426 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.849066 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.861679 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.875271 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:10 crc kubenswrapper[4744]: I1008 09:12:10.885088 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:10Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.034468 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.037273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.037704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.037961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.038415 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.045724 4744 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.046205 4744 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.048111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.048155 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.048168 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.048185 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.048198 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: E1008 09:12:11.065179 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.069410 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.069440 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.069450 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.069469 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.069481 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: E1008 09:12:11.081769 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.086099 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.086304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.086461 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.086548 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.086650 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: E1008 09:12:11.105528 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.113020 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.113177 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.113255 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.113316 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.113388 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: E1008 09:12:11.125871 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.130098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.130127 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.130137 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.130152 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.130163 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: E1008 09:12:11.149285 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: E1008 09:12:11.149433 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.151781 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.151903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.151994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.152107 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.152184 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.254591 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.255115 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.255202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.255272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.255337 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.300688 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.304800 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.309497 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.315739 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.328786 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.341985 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.350324 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.357853 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.357878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.357888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.357902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.357911 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.360022 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.371574 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.385486 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.396975 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.416307 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.432303 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.444335 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.452643 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.452681 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:11 crc kubenswrapper[4744]: E1008 09:12:11.452772 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:11 crc kubenswrapper[4744]: E1008 09:12:11.452969 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.457667 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.464812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.464844 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.464854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.464868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.464878 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.479331 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.492403 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.510587 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.523857 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.535925 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.546028 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.559250 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.567605 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.567656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.567665 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.567679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.567688 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.570973 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.605093 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.640492 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.670522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.670575 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.670591 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.670615 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.670632 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.683432 4744 generic.go:334] "Generic (PLEG): container finished" podID="bfffa6a9-a014-46b8-b4de-3261522cc21e" containerID="fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46" exitCode=0 Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.684019 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" event={"ID":"bfffa6a9-a014-46b8-b4de-3261522cc21e","Type":"ContainerDied","Data":"fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46"} Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.684270 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.696546 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e"} Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.720869 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.760353 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.773724 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.773766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.773777 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.773792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.773805 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.800679 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.844859 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.876306 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.876348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.876359 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.876378 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.876460 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.882028 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.920151 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.960431 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.979369 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.979411 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.979422 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.979438 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:11 crc kubenswrapper[4744]: I1008 09:12:11.979448 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:11Z","lastTransitionTime":"2025-10-08T09:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.001509 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:11Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.041504 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.076222 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.082098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.082126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.082135 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.082151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.082163 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:12Z","lastTransitionTime":"2025-10-08T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.119209 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.165130 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.186047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.186295 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.186454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.186605 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.186716 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:12Z","lastTransitionTime":"2025-10-08T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.202863 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.244534 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.280610 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.289723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.289779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.289793 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.289814 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.289827 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:12Z","lastTransitionTime":"2025-10-08T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.321695 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.360128 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.393289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.393327 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.393341 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.393359 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.393394 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:12Z","lastTransitionTime":"2025-10-08T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.402347 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.452570 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:12 crc kubenswrapper[4744]: E1008 09:12:12.452709 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.497903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.497996 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.498015 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.498694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.498730 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:12Z","lastTransitionTime":"2025-10-08T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.601406 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.601464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.601475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.601494 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.601506 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:12Z","lastTransitionTime":"2025-10-08T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.702447 4744 generic.go:334] "Generic (PLEG): container finished" podID="bfffa6a9-a014-46b8-b4de-3261522cc21e" containerID="4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08" exitCode=0 Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.702509 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" event={"ID":"bfffa6a9-a014-46b8-b4de-3261522cc21e","Type":"ContainerDied","Data":"4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08"} Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.703058 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.703094 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.703108 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.703124 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.703135 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:12Z","lastTransitionTime":"2025-10-08T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.722694 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.747841 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.762052 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.778702 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.791679 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.807969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.808237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.808252 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.808276 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.808289 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:12Z","lastTransitionTime":"2025-10-08T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.810343 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.823425 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.836752 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.854800 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.870616 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.882901 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.898468 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.912691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.912743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.912758 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.912779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.912792 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:12Z","lastTransitionTime":"2025-10-08T09:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.921554 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:12 crc kubenswrapper[4744]: I1008 09:12:12.961096 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:12Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.015888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.015921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.015932 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.015952 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.015966 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:13Z","lastTransitionTime":"2025-10-08T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.118891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.118934 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.118947 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.118965 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.118980 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:13Z","lastTransitionTime":"2025-10-08T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.195839 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.195987 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.196027 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196058 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:12:21.196022617 +0000 UTC m=+36.443667856 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.196112 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196121 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.196152 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196198 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:21.196172762 +0000 UTC m=+36.443818011 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196234 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196272 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196297 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196325 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196274 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196401 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196417 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196383 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:21.196342257 +0000 UTC m=+36.443987496 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196475 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:21.19645928 +0000 UTC m=+36.444104739 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.196498 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:21.196491631 +0000 UTC m=+36.444137140 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.221551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.221600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.221615 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.221638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.221651 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:13Z","lastTransitionTime":"2025-10-08T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.326786 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.326823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.326842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.326865 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.326883 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:13Z","lastTransitionTime":"2025-10-08T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.430332 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.430396 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.430406 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.430432 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.430445 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:13Z","lastTransitionTime":"2025-10-08T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.452765 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.452907 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.454647 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:13 crc kubenswrapper[4744]: E1008 09:12:13.454821 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.539007 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.539081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.539096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.539117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.539131 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:13Z","lastTransitionTime":"2025-10-08T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.642055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.642126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.642138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.642156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.642170 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:13Z","lastTransitionTime":"2025-10-08T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.713864 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" event={"ID":"bfffa6a9-a014-46b8-b4de-3261522cc21e","Type":"ContainerStarted","Data":"fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564"} Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.720890 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3"} Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.722122 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.722212 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.766693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.766936 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.766948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.766968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.766979 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:13Z","lastTransitionTime":"2025-10-08T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.782821 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:13Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.810253 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:13Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.810337 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.811892 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.828322 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:13Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.869960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.870019 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.870035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.870059 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.870074 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:13Z","lastTransitionTime":"2025-10-08T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.872256 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:13Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.888226 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:13Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.905793 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:13Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.918779 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:13Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.931545 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:13Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.944085 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:13Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.957862 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:13Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.972455 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:13Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.978589 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.978656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.978673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.978693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.978709 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:13Z","lastTransitionTime":"2025-10-08T09:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:13 crc kubenswrapper[4744]: I1008 09:12:13.991635 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:13Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.009347 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.023303 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.035085 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.046048 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.056434 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.069038 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.081881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.081950 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.081975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.082006 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.082028 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:14Z","lastTransitionTime":"2025-10-08T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.084008 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.105700 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.121622 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.142000 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.157962 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.184487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.184532 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.184544 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.184561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.184573 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:14Z","lastTransitionTime":"2025-10-08T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.185805 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.203658 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.217987 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.234163 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.249732 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:14Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.288081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.288136 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.288149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.288170 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.288188 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:14Z","lastTransitionTime":"2025-10-08T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.391046 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.391119 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.391150 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.391185 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.391205 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:14Z","lastTransitionTime":"2025-10-08T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.452989 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:14 crc kubenswrapper[4744]: E1008 09:12:14.453172 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.493365 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.493497 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.493515 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.493537 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.493551 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:14Z","lastTransitionTime":"2025-10-08T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.596216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.596263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.596273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.596290 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.596302 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:14Z","lastTransitionTime":"2025-10-08T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.698887 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.698939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.698955 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.698979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.698994 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:14Z","lastTransitionTime":"2025-10-08T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.723766 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.801826 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.801870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.801880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.801899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.801911 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:14Z","lastTransitionTime":"2025-10-08T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.905165 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.905214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.905254 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.905278 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:14 crc kubenswrapper[4744]: I1008 09:12:14.905292 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:14Z","lastTransitionTime":"2025-10-08T09:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.008160 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.008202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.008212 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.008229 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.008241 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:15Z","lastTransitionTime":"2025-10-08T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.110023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.110070 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.110081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.110101 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.110113 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:15Z","lastTransitionTime":"2025-10-08T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.213475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.213525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.213543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.213569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.213586 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:15Z","lastTransitionTime":"2025-10-08T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.317060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.317159 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.317179 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.317203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.317221 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:15Z","lastTransitionTime":"2025-10-08T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.420824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.420880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.420896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.420921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.420933 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:15Z","lastTransitionTime":"2025-10-08T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.452287 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.452357 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:15 crc kubenswrapper[4744]: E1008 09:12:15.452482 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:15 crc kubenswrapper[4744]: E1008 09:12:15.452812 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.464798 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.480257 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.494456 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.511490 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.523325 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.523409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.523425 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.523452 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.523470 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:15Z","lastTransitionTime":"2025-10-08T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.528328 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.547485 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.560520 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.589961 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.612322 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.628367 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.628421 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.628432 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.628448 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.628459 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:15Z","lastTransitionTime":"2025-10-08T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.633185 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.653350 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.666860 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.682448 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.692975 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:15Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.726598 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.730403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.730446 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.730464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.730487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.730506 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:15Z","lastTransitionTime":"2025-10-08T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.834620 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.834698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.834721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.834750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.834771 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:15Z","lastTransitionTime":"2025-10-08T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.939141 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.939769 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.939788 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.939815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:15 crc kubenswrapper[4744]: I1008 09:12:15.939831 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:15Z","lastTransitionTime":"2025-10-08T09:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.046025 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.046111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.046124 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.046141 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.046182 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:16Z","lastTransitionTime":"2025-10-08T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.148310 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.148349 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.148362 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.148406 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.148420 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:16Z","lastTransitionTime":"2025-10-08T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.251911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.251969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.251984 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.252003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.252018 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:16Z","lastTransitionTime":"2025-10-08T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.355502 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.355551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.355560 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.355576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.355588 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:16Z","lastTransitionTime":"2025-10-08T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.453041 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:16 crc kubenswrapper[4744]: E1008 09:12:16.453257 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.459068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.459114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.459133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.459156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.459171 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:16Z","lastTransitionTime":"2025-10-08T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.562770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.562833 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.562845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.562865 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.562884 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:16Z","lastTransitionTime":"2025-10-08T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.666081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.666165 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.666189 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.666222 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.666244 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:16Z","lastTransitionTime":"2025-10-08T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.734058 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/0.log" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.739747 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3" exitCode=1 Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.739835 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3"} Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.740735 4744 scope.go:117] "RemoveContainer" containerID="ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.762910 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:16Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.769306 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.769674 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.769884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.770022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.770137 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:16Z","lastTransitionTime":"2025-10-08T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.801001 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:16Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.815856 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:16Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.828701 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:16Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.855708 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:16Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.873178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.873219 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.873233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.873254 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.873268 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:16Z","lastTransitionTime":"2025-10-08T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.879456 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:16Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.900746 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:16Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.922694 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:16Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.940221 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:16Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.955925 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:16Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.975881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.975938 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.975957 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.975983 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.976000 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:16Z","lastTransitionTime":"2025-10-08T09:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.979732 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:16Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:16 crc kubenswrapper[4744]: I1008 09:12:16.999021 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:16Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.015314 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.042750 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:16Z\\\",\\\"message\\\":\\\"r removal\\\\nI1008 09:12:15.984498 5954 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 09:12:15.984507 5954 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 09:12:15.984533 5954 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 09:12:15.984543 5954 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 09:12:15.984566 5954 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 09:12:15.984614 5954 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 09:12:15.984670 5954 factory.go:656] Stopping watch factory\\\\nI1008 09:12:15.984694 5954 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 09:12:15.984991 5954 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 09:12:15.985012 5954 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 09:12:15.985021 5954 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 09:12:15.985030 5954 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 09:12:15.985038 5954 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 09:12:15.985047 5954 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 09:12:15.985054 5954 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 09:12:15.985072 5954 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.083237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.083298 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.083311 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.083346 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.083358 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:17Z","lastTransitionTime":"2025-10-08T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.187162 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.187225 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.187242 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.187266 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.187284 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:17Z","lastTransitionTime":"2025-10-08T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.291223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.291277 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.291291 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.291313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.291327 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:17Z","lastTransitionTime":"2025-10-08T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.393896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.393940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.393953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.393971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.393984 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:17Z","lastTransitionTime":"2025-10-08T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.452320 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.452331 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:17 crc kubenswrapper[4744]: E1008 09:12:17.452523 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:17 crc kubenswrapper[4744]: E1008 09:12:17.452647 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.496731 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.496776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.496790 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.496808 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.496823 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:17Z","lastTransitionTime":"2025-10-08T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.598889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.598932 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.598944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.598960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.598969 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:17Z","lastTransitionTime":"2025-10-08T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.701583 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.701646 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.701663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.701691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.701713 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:17Z","lastTransitionTime":"2025-10-08T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.745130 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/0.log" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.748467 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197"} Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.748666 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.763407 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.778006 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.794630 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.804599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.804662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.804675 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.804693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.804706 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:17Z","lastTransitionTime":"2025-10-08T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.811797 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.826306 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.841720 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.868736 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.887359 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.907139 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.907545 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.907602 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.907621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.907648 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.907669 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:17Z","lastTransitionTime":"2025-10-08T09:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.927630 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:16Z\\\",\\\"message\\\":\\\"r removal\\\\nI1008 09:12:15.984498 5954 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 09:12:15.984507 5954 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 09:12:15.984533 5954 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 09:12:15.984543 5954 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 09:12:15.984566 5954 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 09:12:15.984614 5954 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 09:12:15.984670 5954 factory.go:656] Stopping watch factory\\\\nI1008 09:12:15.984694 5954 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 09:12:15.984991 5954 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 09:12:15.985012 5954 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 09:12:15.985021 5954 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 09:12:15.985030 5954 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 09:12:15.985038 5954 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 09:12:15.985047 5954 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 09:12:15.985054 5954 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 09:12:15.985072 5954 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.946819 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.959755 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.971851 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:17 crc kubenswrapper[4744]: I1008 09:12:17.986966 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:17Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.011283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.011339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.011351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.011401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.011420 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:18Z","lastTransitionTime":"2025-10-08T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.113704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.113759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.113776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.113796 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.113807 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:18Z","lastTransitionTime":"2025-10-08T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.216683 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.216727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.216745 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.216763 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.216773 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:18Z","lastTransitionTime":"2025-10-08T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.318966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.319026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.319038 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.319053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.319063 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:18Z","lastTransitionTime":"2025-10-08T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.449870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.449927 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.449944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.449969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.449985 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:18Z","lastTransitionTime":"2025-10-08T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.452050 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:18 crc kubenswrapper[4744]: E1008 09:12:18.452196 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.553023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.553102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.553122 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.553151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.553170 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:18Z","lastTransitionTime":"2025-10-08T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.655808 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.655881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.655893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.655912 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.655928 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:18Z","lastTransitionTime":"2025-10-08T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.756103 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/1.log" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.757248 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/0.log" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.758666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.758856 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.758899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.758938 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.758963 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:18Z","lastTransitionTime":"2025-10-08T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.761970 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197" exitCode=1 Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.762031 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197"} Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.762126 4744 scope.go:117] "RemoveContainer" containerID="ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.763882 4744 scope.go:117] "RemoveContainer" containerID="e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197" Oct 08 09:12:18 crc kubenswrapper[4744]: E1008 09:12:18.764457 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.788869 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.813214 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.828410 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.845847 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.861307 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.861336 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.861345 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.861359 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.861388 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:18Z","lastTransitionTime":"2025-10-08T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.861456 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.877241 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.888661 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.901241 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.912574 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.923533 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.935899 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.945743 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.960499 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.964742 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.964774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.964785 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.964801 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.964810 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:18Z","lastTransitionTime":"2025-10-08T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:18 crc kubenswrapper[4744]: I1008 09:12:18.977689 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:16Z\\\",\\\"message\\\":\\\"r removal\\\\nI1008 09:12:15.984498 5954 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 09:12:15.984507 5954 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 09:12:15.984533 5954 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 09:12:15.984543 5954 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 09:12:15.984566 5954 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 09:12:15.984614 5954 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 09:12:15.984670 5954 factory.go:656] Stopping watch factory\\\\nI1008 09:12:15.984694 5954 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 09:12:15.984991 5954 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 09:12:15.985012 5954 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 09:12:15.985021 5954 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 09:12:15.985030 5954 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 09:12:15.985038 5954 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 09:12:15.985047 5954 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 09:12:15.985054 5954 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 09:12:15.985072 5954 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:18Z\\\",\\\"message\\\":\\\"72c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 09:12:17.769575 6092 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 09:12:17.769689 6092 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:18Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.067649 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.067703 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.067725 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.067750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.067768 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:19Z","lastTransitionTime":"2025-10-08T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.103997 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j"] Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.104575 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.106904 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.107092 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.127657 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.143488 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.162281 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e0faba02-dfa7-4a4f-9697-05e0321ed28a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jb62j\" (UID: \"e0faba02-dfa7-4a4f-9697-05e0321ed28a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.162355 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkf9r\" (UniqueName: \"kubernetes.io/projected/e0faba02-dfa7-4a4f-9697-05e0321ed28a-kube-api-access-jkf9r\") pod \"ovnkube-control-plane-749d76644c-jb62j\" (UID: \"e0faba02-dfa7-4a4f-9697-05e0321ed28a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.162580 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e0faba02-dfa7-4a4f-9697-05e0321ed28a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jb62j\" (UID: \"e0faba02-dfa7-4a4f-9697-05e0321ed28a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.162663 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e0faba02-dfa7-4a4f-9697-05e0321ed28a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jb62j\" (UID: \"e0faba02-dfa7-4a4f-9697-05e0321ed28a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.163459 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.172026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.172076 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.172093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.172122 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.172137 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:19Z","lastTransitionTime":"2025-10-08T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.179172 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.197231 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.209954 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.221510 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.236854 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.251786 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.263734 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e0faba02-dfa7-4a4f-9697-05e0321ed28a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jb62j\" (UID: \"e0faba02-dfa7-4a4f-9697-05e0321ed28a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.263791 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e0faba02-dfa7-4a4f-9697-05e0321ed28a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jb62j\" (UID: \"e0faba02-dfa7-4a4f-9697-05e0321ed28a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.263830 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkf9r\" (UniqueName: \"kubernetes.io/projected/e0faba02-dfa7-4a4f-9697-05e0321ed28a-kube-api-access-jkf9r\") pod \"ovnkube-control-plane-749d76644c-jb62j\" (UID: \"e0faba02-dfa7-4a4f-9697-05e0321ed28a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.263904 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e0faba02-dfa7-4a4f-9697-05e0321ed28a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jb62j\" (UID: \"e0faba02-dfa7-4a4f-9697-05e0321ed28a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.265479 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e0faba02-dfa7-4a4f-9697-05e0321ed28a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jb62j\" (UID: \"e0faba02-dfa7-4a4f-9697-05e0321ed28a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.265476 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e0faba02-dfa7-4a4f-9697-05e0321ed28a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jb62j\" (UID: \"e0faba02-dfa7-4a4f-9697-05e0321ed28a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.269558 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.280959 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e0faba02-dfa7-4a4f-9697-05e0321ed28a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jb62j\" (UID: \"e0faba02-dfa7-4a4f-9697-05e0321ed28a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.282740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.282793 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.282806 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.282824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.282835 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:19Z","lastTransitionTime":"2025-10-08T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.284143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkf9r\" (UniqueName: \"kubernetes.io/projected/e0faba02-dfa7-4a4f-9697-05e0321ed28a-kube-api-access-jkf9r\") pod \"ovnkube-control-plane-749d76644c-jb62j\" (UID: \"e0faba02-dfa7-4a4f-9697-05e0321ed28a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.289535 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.304911 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.322635 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.341194 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:16Z\\\",\\\"message\\\":\\\"r removal\\\\nI1008 09:12:15.984498 5954 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 09:12:15.984507 5954 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 09:12:15.984533 5954 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 09:12:15.984543 5954 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 09:12:15.984566 5954 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 09:12:15.984614 5954 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 09:12:15.984670 5954 factory.go:656] Stopping watch factory\\\\nI1008 09:12:15.984694 5954 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 09:12:15.984991 5954 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 09:12:15.985012 5954 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 09:12:15.985021 5954 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 09:12:15.985030 5954 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 09:12:15.985038 5954 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 09:12:15.985047 5954 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 09:12:15.985054 5954 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 09:12:15.985072 5954 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:18Z\\\",\\\"message\\\":\\\"72c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 09:12:17.769575 6092 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 09:12:17.769689 6092 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.354157 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.385959 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.386030 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.386062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.386096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.386121 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:19Z","lastTransitionTime":"2025-10-08T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.430456 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.452418 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.452467 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:19 crc kubenswrapper[4744]: E1008 09:12:19.452656 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:19 crc kubenswrapper[4744]: E1008 09:12:19.452853 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:19 crc kubenswrapper[4744]: W1008 09:12:19.454640 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0faba02_dfa7_4a4f_9697_05e0321ed28a.slice/crio-053018b1300e35332adc21a1e6eca9413bbc505e209ba42209e11e50fea00fb4 WatchSource:0}: Error finding container 053018b1300e35332adc21a1e6eca9413bbc505e209ba42209e11e50fea00fb4: Status 404 returned error can't find the container with id 053018b1300e35332adc21a1e6eca9413bbc505e209ba42209e11e50fea00fb4 Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.490032 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.490083 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.490102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.490127 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.490144 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:19Z","lastTransitionTime":"2025-10-08T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.592340 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.592406 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.592421 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.592439 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.592456 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:19Z","lastTransitionTime":"2025-10-08T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.696810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.696867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.696882 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.696906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.696920 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:19Z","lastTransitionTime":"2025-10-08T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.774551 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/1.log" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.787216 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" event={"ID":"e0faba02-dfa7-4a4f-9697-05e0321ed28a","Type":"ContainerStarted","Data":"4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef"} Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.787304 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" event={"ID":"e0faba02-dfa7-4a4f-9697-05e0321ed28a","Type":"ContainerStarted","Data":"053018b1300e35332adc21a1e6eca9413bbc505e209ba42209e11e50fea00fb4"} Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.799789 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.800044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.800060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.800085 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.800103 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:19Z","lastTransitionTime":"2025-10-08T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.861063 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-cdcbz"] Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.861668 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:19 crc kubenswrapper[4744]: E1008 09:12:19.861750 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.869251 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.869304 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hkbm\" (UniqueName: \"kubernetes.io/projected/be364129-0f0b-41d4-b5f8-1b1def9a1c38-kube-api-access-9hkbm\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.889191 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.904003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.904084 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.904099 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.904122 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.904136 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:19Z","lastTransitionTime":"2025-10-08T09:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.915163 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.939732 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:16Z\\\",\\\"message\\\":\\\"r removal\\\\nI1008 09:12:15.984498 5954 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 09:12:15.984507 5954 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 09:12:15.984533 5954 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 09:12:15.984543 5954 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 09:12:15.984566 5954 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 09:12:15.984614 5954 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 09:12:15.984670 5954 factory.go:656] Stopping watch factory\\\\nI1008 09:12:15.984694 5954 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 09:12:15.984991 5954 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 09:12:15.985012 5954 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 09:12:15.985021 5954 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 09:12:15.985030 5954 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 09:12:15.985038 5954 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 09:12:15.985047 5954 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 09:12:15.985054 5954 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 09:12:15.985072 5954 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:18Z\\\",\\\"message\\\":\\\"72c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 09:12:17.769575 6092 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 09:12:17.769689 6092 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.959886 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.970611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.970659 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hkbm\" (UniqueName: \"kubernetes.io/projected/be364129-0f0b-41d4-b5f8-1b1def9a1c38-kube-api-access-9hkbm\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:19 crc kubenswrapper[4744]: E1008 09:12:19.970808 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:19 crc kubenswrapper[4744]: E1008 09:12:19.970956 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs podName:be364129-0f0b-41d4-b5f8-1b1def9a1c38 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:20.470903537 +0000 UTC m=+35.718548776 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs") pod "network-metrics-daemon-cdcbz" (UID: "be364129-0f0b-41d4-b5f8-1b1def9a1c38") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:19 crc kubenswrapper[4744]: I1008 09:12:19.979600 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.001191 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hkbm\" (UniqueName: \"kubernetes.io/projected/be364129-0f0b-41d4-b5f8-1b1def9a1c38-kube-api-access-9hkbm\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.001229 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:19Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.006667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.006704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.006714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.006731 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.006743 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:20Z","lastTransitionTime":"2025-10-08T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.014921 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.029329 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.044090 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.054139 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.067334 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.081369 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.097059 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.109273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.109326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.109337 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.109359 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.109384 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:20Z","lastTransitionTime":"2025-10-08T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.113120 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.127544 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.144259 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.212412 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.212458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.212470 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.212488 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.212499 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:20Z","lastTransitionTime":"2025-10-08T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.314827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.314872 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.314881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.314900 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.314914 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:20Z","lastTransitionTime":"2025-10-08T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.417389 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.417449 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.417462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.417483 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.417499 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:20Z","lastTransitionTime":"2025-10-08T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.452130 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:20 crc kubenswrapper[4744]: E1008 09:12:20.452308 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.476442 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:20 crc kubenswrapper[4744]: E1008 09:12:20.476625 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:20 crc kubenswrapper[4744]: E1008 09:12:20.476744 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs podName:be364129-0f0b-41d4-b5f8-1b1def9a1c38 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:21.476715366 +0000 UTC m=+36.724360605 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs") pod "network-metrics-daemon-cdcbz" (UID: "be364129-0f0b-41d4-b5f8-1b1def9a1c38") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.520584 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.520645 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.520660 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.520688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.520706 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:20Z","lastTransitionTime":"2025-10-08T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.624232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.624295 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.624309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.624330 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.624344 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:20Z","lastTransitionTime":"2025-10-08T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.728457 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.728505 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.728516 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.728534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.728547 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:20Z","lastTransitionTime":"2025-10-08T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.794938 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" event={"ID":"e0faba02-dfa7-4a4f-9697-05e0321ed28a","Type":"ContainerStarted","Data":"69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037"} Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.822252 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.833305 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.833433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.833453 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.833483 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.833508 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:20Z","lastTransitionTime":"2025-10-08T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.844559 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.866467 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.894525 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.915630 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.936977 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.937582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.937621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.937640 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.937665 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.937684 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:20Z","lastTransitionTime":"2025-10-08T09:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.954940 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.976894 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:20 crc kubenswrapper[4744]: I1008 09:12:20.998226 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:20Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.024916 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:16Z\\\",\\\"message\\\":\\\"r removal\\\\nI1008 09:12:15.984498 5954 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 09:12:15.984507 5954 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 09:12:15.984533 5954 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 09:12:15.984543 5954 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 09:12:15.984566 5954 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 09:12:15.984614 5954 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 09:12:15.984670 5954 factory.go:656] Stopping watch factory\\\\nI1008 09:12:15.984694 5954 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 09:12:15.984991 5954 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 09:12:15.985012 5954 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 09:12:15.985021 5954 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 09:12:15.985030 5954 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 09:12:15.985038 5954 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 09:12:15.985047 5954 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 09:12:15.985054 5954 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 09:12:15.985072 5954 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:18Z\\\",\\\"message\\\":\\\"72c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 09:12:17.769575 6092 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 09:12:17.769689 6092 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:21Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.040972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.041018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.041028 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.041045 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.041058 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.041131 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:21Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.053341 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:21Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.066282 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:21Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.088779 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:21Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.111815 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:21Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.132826 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:21Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.144039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.144090 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.144102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.144123 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.144134 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.247955 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.248024 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.248040 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.248071 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.248091 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.285754 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.285910 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.285944 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.285982 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:12:37.285942827 +0000 UTC m=+52.533588076 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.286092 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.286139 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.286152 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.286181 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.286209 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:37.286191995 +0000 UTC m=+52.533837234 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.286255 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.286263 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.286411 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:37.28635447 +0000 UTC m=+52.533999979 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.286448 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.286515 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:37.286502945 +0000 UTC m=+52.534148194 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.286710 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.286775 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.286802 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.286924 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:37.286888866 +0000 UTC m=+52.534534145 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.351891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.351963 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.351983 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.352005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.352016 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.452979 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.453013 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.453313 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.453357 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.453591 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.454198 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.455753 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.455806 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.455825 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.455849 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.455871 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.486923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.487018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.487036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.487059 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.487099 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.488136 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.488457 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.488591 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs podName:be364129-0f0b-41d4-b5f8-1b1def9a1c38 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:23.488560213 +0000 UTC m=+38.736205642 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs") pod "network-metrics-daemon-cdcbz" (UID: "be364129-0f0b-41d4-b5f8-1b1def9a1c38") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.509595 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:21Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.515957 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.516031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.516049 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.516095 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.516115 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.539263 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:21Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.544878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.544948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.544971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.545003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.545026 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.568633 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:21Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.574419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.574491 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.574519 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.574555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.574579 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.593210 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:21Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.599648 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.599728 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.599795 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.599830 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.599855 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.616533 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:21Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:21 crc kubenswrapper[4744]: E1008 09:12:21.616756 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.619026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.619081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.619099 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.619124 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.619146 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.723220 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.723269 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.723288 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.723315 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.723332 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.826303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.826429 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.826463 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.826498 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.826522 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.930402 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.930467 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.930490 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.930520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:21 crc kubenswrapper[4744]: I1008 09:12:21.930534 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:21Z","lastTransitionTime":"2025-10-08T09:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.035143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.035217 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.035237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.035263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.035281 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:22Z","lastTransitionTime":"2025-10-08T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.139250 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.139330 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.139351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.139426 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.139455 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:22Z","lastTransitionTime":"2025-10-08T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.243920 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.244000 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.244022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.244055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.244100 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:22Z","lastTransitionTime":"2025-10-08T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.350110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.350180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.350199 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.350241 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.350262 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:22Z","lastTransitionTime":"2025-10-08T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.452242 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:22 crc kubenswrapper[4744]: E1008 09:12:22.452548 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.458962 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.459038 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.459061 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.459090 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.459114 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:22Z","lastTransitionTime":"2025-10-08T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.562983 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.563570 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.563847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.564029 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.564186 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:22Z","lastTransitionTime":"2025-10-08T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.668216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.668272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.668285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.668304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.668319 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:22Z","lastTransitionTime":"2025-10-08T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.772638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.772704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.772729 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.772763 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.772788 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:22Z","lastTransitionTime":"2025-10-08T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.877332 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.877422 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.877442 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.877468 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.877488 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:22Z","lastTransitionTime":"2025-10-08T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.981864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.981941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.981960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.981987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:22 crc kubenswrapper[4744]: I1008 09:12:22.982006 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:22Z","lastTransitionTime":"2025-10-08T09:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.081240 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.086276 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.086606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.086794 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.087024 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.087198 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:23Z","lastTransitionTime":"2025-10-08T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.109242 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.132062 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.151795 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.168291 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.188030 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.189875 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.190004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.190900 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.191050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.191147 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:23Z","lastTransitionTime":"2025-10-08T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.237298 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.254026 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.266045 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.279130 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.291410 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.293915 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.293950 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.293967 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.293989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.294001 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:23Z","lastTransitionTime":"2025-10-08T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.308181 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.323673 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.337785 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.357162 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:16Z\\\",\\\"message\\\":\\\"r removal\\\\nI1008 09:12:15.984498 5954 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 09:12:15.984507 5954 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 09:12:15.984533 5954 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 09:12:15.984543 5954 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 09:12:15.984566 5954 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 09:12:15.984614 5954 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 09:12:15.984670 5954 factory.go:656] Stopping watch factory\\\\nI1008 09:12:15.984694 5954 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 09:12:15.984991 5954 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 09:12:15.985012 5954 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 09:12:15.985021 5954 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 09:12:15.985030 5954 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 09:12:15.985038 5954 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 09:12:15.985047 5954 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 09:12:15.985054 5954 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 09:12:15.985072 5954 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:18Z\\\",\\\"message\\\":\\\"72c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 09:12:17.769575 6092 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 09:12:17.769689 6092 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.370362 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.383513 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:23Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.396645 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.396695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.396707 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.396726 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.396735 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:23Z","lastTransitionTime":"2025-10-08T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.452358 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.452494 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.452530 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:23 crc kubenswrapper[4744]: E1008 09:12:23.453069 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:23 crc kubenswrapper[4744]: E1008 09:12:23.453295 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:23 crc kubenswrapper[4744]: E1008 09:12:23.453574 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.501638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.501708 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.501732 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.501763 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.501784 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:23Z","lastTransitionTime":"2025-10-08T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.511434 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:23 crc kubenswrapper[4744]: E1008 09:12:23.511595 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:23 crc kubenswrapper[4744]: E1008 09:12:23.511664 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs podName:be364129-0f0b-41d4-b5f8-1b1def9a1c38 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:27.511645775 +0000 UTC m=+42.759291014 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs") pod "network-metrics-daemon-cdcbz" (UID: "be364129-0f0b-41d4-b5f8-1b1def9a1c38") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.606025 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.606116 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.606143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.606177 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.606204 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:23Z","lastTransitionTime":"2025-10-08T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.709144 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.709190 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.709201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.709218 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.709230 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:23Z","lastTransitionTime":"2025-10-08T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.811890 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.811958 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.811972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.811995 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.812010 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:23Z","lastTransitionTime":"2025-10-08T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.915435 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.916496 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.916530 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.916563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:23 crc kubenswrapper[4744]: I1008 09:12:23.916585 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:23Z","lastTransitionTime":"2025-10-08T09:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.019465 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.019516 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.019528 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.019547 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.019558 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:24Z","lastTransitionTime":"2025-10-08T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.122919 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.122996 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.123017 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.123045 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.123064 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:24Z","lastTransitionTime":"2025-10-08T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.226233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.226313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.226331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.226354 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.226394 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:24Z","lastTransitionTime":"2025-10-08T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.330249 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.330317 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.330397 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.330438 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.330459 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:24Z","lastTransitionTime":"2025-10-08T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.434068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.434127 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.434143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.434165 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.434181 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:24Z","lastTransitionTime":"2025-10-08T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.452566 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:24 crc kubenswrapper[4744]: E1008 09:12:24.452749 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.537231 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.537292 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.537309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.537328 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.537341 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:24Z","lastTransitionTime":"2025-10-08T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.640834 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.640893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.640906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.640928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.640942 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:24Z","lastTransitionTime":"2025-10-08T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.744058 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.744125 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.744148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.744178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.744203 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:24Z","lastTransitionTime":"2025-10-08T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.847880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.847945 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.847969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.847998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.848023 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:24Z","lastTransitionTime":"2025-10-08T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.951723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.951799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.951821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.951849 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:24 crc kubenswrapper[4744]: I1008 09:12:24.951872 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:24Z","lastTransitionTime":"2025-10-08T09:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.054810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.054856 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.054870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.054889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.054902 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:25Z","lastTransitionTime":"2025-10-08T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.158500 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.158547 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.158561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.158581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.158596 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:25Z","lastTransitionTime":"2025-10-08T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.261592 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.261657 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.261673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.261698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.261718 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:25Z","lastTransitionTime":"2025-10-08T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.365643 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.365694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.365708 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.365728 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.365742 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:25Z","lastTransitionTime":"2025-10-08T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.452418 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.452472 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.452581 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:25 crc kubenswrapper[4744]: E1008 09:12:25.452615 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:25 crc kubenswrapper[4744]: E1008 09:12:25.452804 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:25 crc kubenswrapper[4744]: E1008 09:12:25.453082 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.470209 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.470279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.470300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.470341 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.470362 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:25Z","lastTransitionTime":"2025-10-08T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.480775 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.502943 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.521538 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.539242 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.561815 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.574875 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.574928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.574943 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.574964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.574976 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:25Z","lastTransitionTime":"2025-10-08T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.581780 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.596586 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.614992 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.633255 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.650870 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.665606 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.679524 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.679886 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.679918 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.679928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.679943 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.679954 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:25Z","lastTransitionTime":"2025-10-08T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.693165 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.711121 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ebb00bfd89a34f6b2409b79141ad0cbe13d7292977a18f08bc8a3544c94acca3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:16Z\\\",\\\"message\\\":\\\"r removal\\\\nI1008 09:12:15.984498 5954 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 09:12:15.984507 5954 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 09:12:15.984533 5954 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 09:12:15.984543 5954 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 09:12:15.984566 5954 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 09:12:15.984614 5954 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 09:12:15.984670 5954 factory.go:656] Stopping watch factory\\\\nI1008 09:12:15.984694 5954 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 09:12:15.984991 5954 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 09:12:15.985012 5954 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 09:12:15.985021 5954 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 09:12:15.985030 5954 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 09:12:15.985038 5954 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 09:12:15.985047 5954 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 09:12:15.985054 5954 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 09:12:15.985072 5954 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:18Z\\\",\\\"message\\\":\\\"72c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 09:12:17.769575 6092 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 09:12:17.769689 6092 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.724293 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.733877 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:25Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.783141 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.783298 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.783381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.783453 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.783509 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:25Z","lastTransitionTime":"2025-10-08T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.886419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.886475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.886491 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.886515 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.886534 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:25Z","lastTransitionTime":"2025-10-08T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.989875 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.989979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.990002 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.990031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:25 crc kubenswrapper[4744]: I1008 09:12:25.990051 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:25Z","lastTransitionTime":"2025-10-08T09:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.094041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.094112 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.094134 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.094202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.094224 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:26Z","lastTransitionTime":"2025-10-08T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.198456 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.198561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.198585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.198628 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.198650 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:26Z","lastTransitionTime":"2025-10-08T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.302655 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.302727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.302747 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.302773 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.302792 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:26Z","lastTransitionTime":"2025-10-08T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.406451 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.406525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.406544 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.406571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.406591 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:26Z","lastTransitionTime":"2025-10-08T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.453194 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:26 crc kubenswrapper[4744]: E1008 09:12:26.453627 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.510227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.510294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.510317 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.510348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.510367 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:26Z","lastTransitionTime":"2025-10-08T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.613927 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.613970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.613983 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.614001 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.614014 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:26Z","lastTransitionTime":"2025-10-08T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.718043 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.719951 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.720214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.720441 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.721351 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:26Z","lastTransitionTime":"2025-10-08T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.825148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.825561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.825586 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.825623 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.825649 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:26Z","lastTransitionTime":"2025-10-08T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.930143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.930233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.930260 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.930289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:26 crc kubenswrapper[4744]: I1008 09:12:26.930310 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:26Z","lastTransitionTime":"2025-10-08T09:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.033649 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.033725 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.033751 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.033785 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.033811 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:27Z","lastTransitionTime":"2025-10-08T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.137861 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.138274 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.138480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.138658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.138810 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:27Z","lastTransitionTime":"2025-10-08T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.243033 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.243102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.243123 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.243148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.243169 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:27Z","lastTransitionTime":"2025-10-08T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.347050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.347106 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.347120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.347142 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.347154 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:27Z","lastTransitionTime":"2025-10-08T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.449632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.449675 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.449687 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.449704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.449716 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:27Z","lastTransitionTime":"2025-10-08T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.452427 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.452449 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:27 crc kubenswrapper[4744]: E1008 09:12:27.452558 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:27 crc kubenswrapper[4744]: E1008 09:12:27.452781 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.452428 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:27 crc kubenswrapper[4744]: E1008 09:12:27.453760 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.552510 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.552862 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.553009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.553122 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.553205 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:27Z","lastTransitionTime":"2025-10-08T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.563296 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:27 crc kubenswrapper[4744]: E1008 09:12:27.563552 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:27 crc kubenswrapper[4744]: E1008 09:12:27.563689 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs podName:be364129-0f0b-41d4-b5f8-1b1def9a1c38 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:35.563661187 +0000 UTC m=+50.811306446 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs") pod "network-metrics-daemon-cdcbz" (UID: "be364129-0f0b-41d4-b5f8-1b1def9a1c38") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.655302 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.655613 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.655723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.655801 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.655881 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:27Z","lastTransitionTime":"2025-10-08T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.758953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.759036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.759060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.759091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.759119 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:27Z","lastTransitionTime":"2025-10-08T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.862515 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.862600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.862622 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.862649 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.862675 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:27Z","lastTransitionTime":"2025-10-08T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.966399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.966440 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.966451 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.966467 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:27 crc kubenswrapper[4744]: I1008 09:12:27.966478 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:27Z","lastTransitionTime":"2025-10-08T09:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.069123 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.069205 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.069228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.069255 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.069273 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:28Z","lastTransitionTime":"2025-10-08T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.171902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.171944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.171958 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.171978 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.172014 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:28Z","lastTransitionTime":"2025-10-08T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.274903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.274991 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.275012 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.275044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.275063 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:28Z","lastTransitionTime":"2025-10-08T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.378848 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.378928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.378950 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.378984 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.379010 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:28Z","lastTransitionTime":"2025-10-08T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.453081 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:28 crc kubenswrapper[4744]: E1008 09:12:28.453366 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.484297 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.484426 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.484460 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.484495 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.484523 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:28Z","lastTransitionTime":"2025-10-08T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.588486 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.588574 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.588600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.588636 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.588662 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:28Z","lastTransitionTime":"2025-10-08T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.691559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.691612 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.691624 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.691640 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.691652 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:28Z","lastTransitionTime":"2025-10-08T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.794166 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.794230 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.794248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.794275 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.794293 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:28Z","lastTransitionTime":"2025-10-08T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.898105 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.898195 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.898217 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.898250 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:28 crc kubenswrapper[4744]: I1008 09:12:28.898272 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:28Z","lastTransitionTime":"2025-10-08T09:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.000815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.000877 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.000895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.000917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.000935 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:29Z","lastTransitionTime":"2025-10-08T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.104894 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.104970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.104992 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.105023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.105047 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:29Z","lastTransitionTime":"2025-10-08T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.208273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.208351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.208398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.208428 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.208444 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:29Z","lastTransitionTime":"2025-10-08T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.311459 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.311510 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.311534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.311553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.311566 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:29Z","lastTransitionTime":"2025-10-08T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.414302 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.414385 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.414401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.414430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.414444 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:29Z","lastTransitionTime":"2025-10-08T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.452546 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.452638 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.452571 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:29 crc kubenswrapper[4744]: E1008 09:12:29.452738 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:29 crc kubenswrapper[4744]: E1008 09:12:29.452864 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:29 crc kubenswrapper[4744]: E1008 09:12:29.452997 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.518417 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.518491 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.518512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.518541 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.518564 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:29Z","lastTransitionTime":"2025-10-08T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.622845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.622901 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.622941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.622960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.622973 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:29Z","lastTransitionTime":"2025-10-08T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.727055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.727114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.727128 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.727149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.727164 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:29Z","lastTransitionTime":"2025-10-08T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.830961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.831021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.831034 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.831053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.831066 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:29Z","lastTransitionTime":"2025-10-08T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.934606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.934670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.934685 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.934709 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:29 crc kubenswrapper[4744]: I1008 09:12:29.934725 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:29Z","lastTransitionTime":"2025-10-08T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.037488 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.037556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.037580 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.037610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.037632 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:30Z","lastTransitionTime":"2025-10-08T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.141949 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.142022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.142047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.142077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.142096 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:30Z","lastTransitionTime":"2025-10-08T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.246085 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.246161 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.246181 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.246215 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.246236 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:30Z","lastTransitionTime":"2025-10-08T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.350044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.350097 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.350117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.350142 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.350160 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:30Z","lastTransitionTime":"2025-10-08T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.452930 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:30 crc kubenswrapper[4744]: E1008 09:12:30.453194 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.453765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.453808 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.453828 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.453853 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.453874 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:30Z","lastTransitionTime":"2025-10-08T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.557912 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.557989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.558008 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.558035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.558056 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:30Z","lastTransitionTime":"2025-10-08T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.661523 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.661559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.661569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.661583 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.661593 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:30Z","lastTransitionTime":"2025-10-08T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.765545 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.765597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.765614 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.765638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.765656 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:30Z","lastTransitionTime":"2025-10-08T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.867990 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.868040 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.868051 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.868073 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.868084 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:30Z","lastTransitionTime":"2025-10-08T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.971970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.972069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.972088 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.972116 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:30 crc kubenswrapper[4744]: I1008 09:12:30.972137 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:30Z","lastTransitionTime":"2025-10-08T09:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.076566 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.076631 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.076651 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.076693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.076710 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.180938 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.181002 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.181016 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.181043 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.181062 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.285793 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.285871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.285890 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.285917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.285935 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.344468 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.346614 4744 scope.go:117] "RemoveContainer" containerID="e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.371779 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.389643 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.389713 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.389739 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.389772 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.389795 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.404241 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.440199 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:18Z\\\",\\\"message\\\":\\\"72c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 09:12:17.769575 6092 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 09:12:17.769689 6092 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.452257 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.452329 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:31 crc kubenswrapper[4744]: E1008 09:12:31.452421 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:31 crc kubenswrapper[4744]: E1008 09:12:31.452540 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.452149 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:31 crc kubenswrapper[4744]: E1008 09:12:31.452750 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.468244 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.482080 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.493360 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.493417 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.493432 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.493454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.493469 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.500124 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.514840 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.528258 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.539548 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.553083 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.568073 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.582011 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.596470 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.597479 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.597534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.597551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.597575 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.597591 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.609039 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.621868 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.626383 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.626406 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.626417 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.626432 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.626444 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.635628 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: E1008 09:12:31.638994 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.642553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.642582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.642595 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.642615 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.642630 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: E1008 09:12:31.657712 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.662216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.662271 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.662286 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.662307 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.662322 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: E1008 09:12:31.676604 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.681500 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.681548 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.681563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.681582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.681593 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: E1008 09:12:31.693929 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.697634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.697688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.697701 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.697719 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.697730 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: E1008 09:12:31.711715 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: E1008 09:12:31.711840 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.713783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.713818 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.713832 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.713853 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.713863 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.816978 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.817044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.817057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.817081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.817095 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.844676 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/1.log" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.849708 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0"} Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.850618 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.873500 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.896851 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.921282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.921326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.921340 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.921361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.921400 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:31Z","lastTransitionTime":"2025-10-08T09:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.931924 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:18Z\\\",\\\"message\\\":\\\"72c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 09:12:17.769575 6092 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 09:12:17.769689 6092 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.956814 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.972279 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:31 crc kubenswrapper[4744]: I1008 09:12:31.986967 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.000090 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:31Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.014927 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:32Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.023543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.023578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.023591 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.023608 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.023619 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:32Z","lastTransitionTime":"2025-10-08T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.031279 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:32Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.047138 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:32Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.064146 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:32Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.078714 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:32Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.093043 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:32Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.110221 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:32Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.125891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.125941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.125953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.125972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.125986 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:32Z","lastTransitionTime":"2025-10-08T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.126681 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:32Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.139949 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:32Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.228785 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.228842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.228859 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.228880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.228896 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:32Z","lastTransitionTime":"2025-10-08T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.331419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.331484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.331497 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.331517 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.331536 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:32Z","lastTransitionTime":"2025-10-08T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.434867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.434919 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.434931 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.434947 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.434959 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:32Z","lastTransitionTime":"2025-10-08T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.452335 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:32 crc kubenswrapper[4744]: E1008 09:12:32.452566 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.538520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.538581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.538600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.538627 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.538648 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:32Z","lastTransitionTime":"2025-10-08T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.642201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.642272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.642291 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.642319 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.642338 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:32Z","lastTransitionTime":"2025-10-08T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.745883 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.745932 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.745944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.745961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.745974 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:32Z","lastTransitionTime":"2025-10-08T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.849583 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.849651 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.849676 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.849724 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.849748 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:32Z","lastTransitionTime":"2025-10-08T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.854914 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/2.log" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.855881 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/1.log" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.859524 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0" exitCode=1 Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.859561 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0"} Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.859603 4744 scope.go:117] "RemoveContainer" containerID="e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.860695 4744 scope.go:117] "RemoveContainer" containerID="8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0" Oct 08 09:12:32 crc kubenswrapper[4744]: E1008 09:12:32.861001 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.891145 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:32Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.913742 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:32Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.953305 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.953399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.953419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.953442 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.953462 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:32Z","lastTransitionTime":"2025-10-08T09:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:32 crc kubenswrapper[4744]: I1008 09:12:32.973955 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e9305600034abe330874d63e1c532199cb79dbbd0eb60f233a662a1f340d3197\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:18Z\\\",\\\"message\\\":\\\"72c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 09:12:17.769575 6092 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 09:12:17.769689 6092 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:32Z\\\",\\\"message\\\":\\\"o:386] Retry successful for *v1.Pod openshift-multus/multus-dldlq after 0 failed attempt(s)\\\\nI1008 09:12:32.351708 6275 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 09:12:32.351717 6275 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j\\\\nI1008 09:12:32.351729 6275 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 09:12:32.351734 6275 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j in node crc\\\\nI1008 09:12:32.351704 6275 event.go:377] Event(v1.ObjectReference{Kind:\\\\\\\"Pod\\\\\\\", Namespace:\\\\\\\"openshift-multus\\\\\\\", Name:\\\\\\\"network-metrics-daemon-cdcbz\\\\\\\", UID:\\\\\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26893\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:32Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.009246 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.030263 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.043547 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.055553 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.055761 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.055779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.055787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.055800 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.055809 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:33Z","lastTransitionTime":"2025-10-08T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.073839 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.087174 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.102122 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.118189 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.130454 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.145423 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.158831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.158861 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.158870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.158883 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.158893 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:33Z","lastTransitionTime":"2025-10-08T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.160539 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.175846 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.189355 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.273395 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.273447 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.273457 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.273473 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.273485 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:33Z","lastTransitionTime":"2025-10-08T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.376664 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.376733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.376751 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.376776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.376794 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:33Z","lastTransitionTime":"2025-10-08T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.452539 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.452597 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:33 crc kubenswrapper[4744]: E1008 09:12:33.452735 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.452868 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:33 crc kubenswrapper[4744]: E1008 09:12:33.452853 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:33 crc kubenswrapper[4744]: E1008 09:12:33.452951 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.480005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.480065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.480081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.480104 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.480122 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:33Z","lastTransitionTime":"2025-10-08T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.583472 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.583543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.583567 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.583598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.583623 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:33Z","lastTransitionTime":"2025-10-08T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.686769 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.686858 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.686882 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.686914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.686936 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:33Z","lastTransitionTime":"2025-10-08T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.790097 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.790178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.790203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.790239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.790257 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:33Z","lastTransitionTime":"2025-10-08T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.866224 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/2.log" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.875641 4744 scope.go:117] "RemoveContainer" containerID="8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0" Oct 08 09:12:33 crc kubenswrapper[4744]: E1008 09:12:33.876095 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.893456 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.893522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.893532 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.893549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.893561 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:33Z","lastTransitionTime":"2025-10-08T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.902326 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.918018 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.933690 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.945648 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.957127 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.970787 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.986797 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.996658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.997061 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.997224 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.997325 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:33 crc kubenswrapper[4744]: I1008 09:12:33.997495 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:33Z","lastTransitionTime":"2025-10-08T09:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.001947 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:33Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.018738 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:34Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.034102 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:34Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.050082 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:34Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.065563 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:34Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.084416 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:34Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.100553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.100590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.100599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.100613 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.100623 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:34Z","lastTransitionTime":"2025-10-08T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.108305 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:32Z\\\",\\\"message\\\":\\\"o:386] Retry successful for *v1.Pod openshift-multus/multus-dldlq after 0 failed attempt(s)\\\\nI1008 09:12:32.351708 6275 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 09:12:32.351717 6275 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j\\\\nI1008 09:12:32.351729 6275 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 09:12:32.351734 6275 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j in node crc\\\\nI1008 09:12:32.351704 6275 event.go:377] Event(v1.ObjectReference{Kind:\\\\\\\"Pod\\\\\\\", Namespace:\\\\\\\"openshift-multus\\\\\\\", Name:\\\\\\\"network-metrics-daemon-cdcbz\\\\\\\", UID:\\\\\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26893\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:34Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.121666 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:34Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.132801 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:34Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.202817 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.202847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.202856 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.202871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.202882 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:34Z","lastTransitionTime":"2025-10-08T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.305269 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.305307 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.305317 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.305331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.305342 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:34Z","lastTransitionTime":"2025-10-08T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.407555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.407654 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.407673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.407698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.407717 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:34Z","lastTransitionTime":"2025-10-08T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.452821 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:34 crc kubenswrapper[4744]: E1008 09:12:34.452980 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.510203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.510258 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.510267 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.510280 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.510290 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:34Z","lastTransitionTime":"2025-10-08T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.614767 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.614845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.614867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.614896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.614978 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:34Z","lastTransitionTime":"2025-10-08T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.719562 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.719632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.719647 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.719670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.719684 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:34Z","lastTransitionTime":"2025-10-08T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.822982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.823032 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.823042 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.823058 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.823070 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:34Z","lastTransitionTime":"2025-10-08T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.925581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.925629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.925642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.925660 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:34 crc kubenswrapper[4744]: I1008 09:12:34.925671 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:34Z","lastTransitionTime":"2025-10-08T09:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.029325 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.029398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.029409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.029430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.029440 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:35Z","lastTransitionTime":"2025-10-08T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.132725 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.132798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.132810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.132831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.132846 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:35Z","lastTransitionTime":"2025-10-08T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.235515 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.235563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.235573 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.235592 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.235604 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:35Z","lastTransitionTime":"2025-10-08T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.339908 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.339987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.340018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.340050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.340070 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:35Z","lastTransitionTime":"2025-10-08T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.443143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.443197 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.443213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.443239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.443266 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:35Z","lastTransitionTime":"2025-10-08T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.452647 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:35 crc kubenswrapper[4744]: E1008 09:12:35.452830 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.453317 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:35 crc kubenswrapper[4744]: E1008 09:12:35.453421 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.453532 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:35 crc kubenswrapper[4744]: E1008 09:12:35.453653 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.479271 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:32Z\\\",\\\"message\\\":\\\"o:386] Retry successful for *v1.Pod openshift-multus/multus-dldlq after 0 failed attempt(s)\\\\nI1008 09:12:32.351708 6275 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 09:12:32.351717 6275 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j\\\\nI1008 09:12:32.351729 6275 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 09:12:32.351734 6275 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j in node crc\\\\nI1008 09:12:32.351704 6275 event.go:377] Event(v1.ObjectReference{Kind:\\\\\\\"Pod\\\\\\\", Namespace:\\\\\\\"openshift-multus\\\\\\\", Name:\\\\\\\"network-metrics-daemon-cdcbz\\\\\\\", UID:\\\\\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26893\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.494740 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.510948 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.530070 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.545098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.545148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.545159 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.545173 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.545185 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:35Z","lastTransitionTime":"2025-10-08T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.551293 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.571519 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.591296 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.591987 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:35 crc kubenswrapper[4744]: E1008 09:12:35.592157 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:35 crc kubenswrapper[4744]: E1008 09:12:35.592237 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs podName:be364129-0f0b-41d4-b5f8-1b1def9a1c38 nodeName:}" failed. No retries permitted until 2025-10-08 09:12:51.592208372 +0000 UTC m=+66.839853631 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs") pod "network-metrics-daemon-cdcbz" (UID: "be364129-0f0b-41d4-b5f8-1b1def9a1c38") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.612135 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.628745 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.639890 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.648590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.648634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.648650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.648671 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.648685 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:35Z","lastTransitionTime":"2025-10-08T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.656955 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.672205 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.695749 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.715607 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.734854 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.751899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.751966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.751995 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.752026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.752104 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:35Z","lastTransitionTime":"2025-10-08T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.755085 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:35Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.855598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.855655 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.855666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.855683 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.855718 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:35Z","lastTransitionTime":"2025-10-08T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.959513 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.959574 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.959592 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.959613 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:35 crc kubenswrapper[4744]: I1008 09:12:35.959626 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:35Z","lastTransitionTime":"2025-10-08T09:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.063564 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.063641 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.063658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.063683 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.063702 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:36Z","lastTransitionTime":"2025-10-08T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.167084 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.167133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.167145 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.167163 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.167177 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:36Z","lastTransitionTime":"2025-10-08T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.270348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.270723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.270817 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.270911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.271042 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:36Z","lastTransitionTime":"2025-10-08T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.374612 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.374663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.374680 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.374703 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.374720 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:36Z","lastTransitionTime":"2025-10-08T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.452345 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:36 crc kubenswrapper[4744]: E1008 09:12:36.452618 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.477995 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.478040 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.478059 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.478082 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.478103 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:36Z","lastTransitionTime":"2025-10-08T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.580849 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.580896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.580909 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.580929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.580942 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:36Z","lastTransitionTime":"2025-10-08T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.685835 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.685908 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.685926 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.685958 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.685977 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:36Z","lastTransitionTime":"2025-10-08T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.788998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.789093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.789115 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.789142 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.789162 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:36Z","lastTransitionTime":"2025-10-08T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.892546 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.892611 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.892630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.892656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.892676 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:36Z","lastTransitionTime":"2025-10-08T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.995888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.995937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.995950 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.995966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:36 crc kubenswrapper[4744]: I1008 09:12:36.995979 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:36Z","lastTransitionTime":"2025-10-08T09:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.098886 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.098994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.099082 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.099106 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.099129 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:37Z","lastTransitionTime":"2025-10-08T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.202338 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.202444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.202470 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.202501 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.202527 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:37Z","lastTransitionTime":"2025-10-08T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.307401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.307477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.307499 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.307526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.307545 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:37Z","lastTransitionTime":"2025-10-08T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.312228 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.312484 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:13:09.31245157 +0000 UTC m=+84.560096829 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.312537 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.312586 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.312620 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.312643 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.312762 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.312792 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.312827 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.312848 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.312885 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.312927 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.312955 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.312808 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:13:09.3128007 +0000 UTC m=+84.560445939 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.312759 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.313048 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 09:13:09.313020528 +0000 UTC m=+84.560665777 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.313154 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 09:13:09.313127971 +0000 UTC m=+84.560773260 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.313193 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:13:09.313177143 +0000 UTC m=+84.560822422 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.410785 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.410834 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.410846 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.410863 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.410876 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:37Z","lastTransitionTime":"2025-10-08T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.452730 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.452784 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.452804 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.452899 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.452984 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:37 crc kubenswrapper[4744]: E1008 09:12:37.453045 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.514091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.514127 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.514138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.514153 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.514163 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:37Z","lastTransitionTime":"2025-10-08T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.616529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.616576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.616588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.616604 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.616616 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:37Z","lastTransitionTime":"2025-10-08T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.719072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.719120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.719138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.719156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.719172 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:37Z","lastTransitionTime":"2025-10-08T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.822165 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.822214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.822232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.822248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.822259 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:37Z","lastTransitionTime":"2025-10-08T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.925832 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.925896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.925917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.925944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:37 crc kubenswrapper[4744]: I1008 09:12:37.925965 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:37Z","lastTransitionTime":"2025-10-08T09:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.029246 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.029285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.029294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.029309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.029320 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:38Z","lastTransitionTime":"2025-10-08T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.135439 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.135505 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.135519 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.135540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.135554 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:38Z","lastTransitionTime":"2025-10-08T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.239915 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.239968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.239987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.240010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.240028 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:38Z","lastTransitionTime":"2025-10-08T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.343247 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.343306 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.343320 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.343342 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.343358 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:38Z","lastTransitionTime":"2025-10-08T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.446040 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.446069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.446079 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.446093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.446105 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:38Z","lastTransitionTime":"2025-10-08T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.452838 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:38 crc kubenswrapper[4744]: E1008 09:12:38.452993 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.550940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.550990 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.551002 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.551022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.551034 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:38Z","lastTransitionTime":"2025-10-08T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.653890 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.653940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.653949 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.653970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.653982 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:38Z","lastTransitionTime":"2025-10-08T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.756315 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.756345 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.756354 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.756380 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.756391 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:38Z","lastTransitionTime":"2025-10-08T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.860273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.860344 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.860399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.860426 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.860446 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:38Z","lastTransitionTime":"2025-10-08T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.962956 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.962986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.962994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.963010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:38 crc kubenswrapper[4744]: I1008 09:12:38.963021 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:38Z","lastTransitionTime":"2025-10-08T09:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.065303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.065363 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.065408 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.065429 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.065443 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:39Z","lastTransitionTime":"2025-10-08T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.137223 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.155845 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.162336 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.168247 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.168316 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.168343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.168407 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.168431 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:39Z","lastTransitionTime":"2025-10-08T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.181215 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.200848 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.217466 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.231822 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.244806 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.257264 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.271332 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.271393 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.271407 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.271427 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.271440 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:39Z","lastTransitionTime":"2025-10-08T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.275293 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.295904 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.317597 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:32Z\\\",\\\"message\\\":\\\"o:386] Retry successful for *v1.Pod openshift-multus/multus-dldlq after 0 failed attempt(s)\\\\nI1008 09:12:32.351708 6275 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 09:12:32.351717 6275 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j\\\\nI1008 09:12:32.351729 6275 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 09:12:32.351734 6275 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j in node crc\\\\nI1008 09:12:32.351704 6275 event.go:377] Event(v1.ObjectReference{Kind:\\\\\\\"Pod\\\\\\\", Namespace:\\\\\\\"openshift-multus\\\\\\\", Name:\\\\\\\"network-metrics-daemon-cdcbz\\\\\\\", UID:\\\\\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26893\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.332735 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.345710 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.360294 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.374325 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.374400 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.374419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.374435 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.374445 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:39Z","lastTransitionTime":"2025-10-08T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.375290 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.390783 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.405116 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:39Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.452436 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:39 crc kubenswrapper[4744]: E1008 09:12:39.452614 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.452724 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.452806 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:39 crc kubenswrapper[4744]: E1008 09:12:39.452993 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:39 crc kubenswrapper[4744]: E1008 09:12:39.453273 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.478073 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.478178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.478223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.478247 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.478265 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:39Z","lastTransitionTime":"2025-10-08T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.587529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.587593 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.587609 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.587662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.587750 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:39Z","lastTransitionTime":"2025-10-08T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.692171 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.692234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.692245 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.692260 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.692271 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:39Z","lastTransitionTime":"2025-10-08T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.795476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.795540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.795559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.795584 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.795602 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:39Z","lastTransitionTime":"2025-10-08T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.899755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.899833 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.899857 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.899888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:39 crc kubenswrapper[4744]: I1008 09:12:39.899911 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:39Z","lastTransitionTime":"2025-10-08T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.003131 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.003200 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.003213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.003229 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.003243 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:40Z","lastTransitionTime":"2025-10-08T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.106088 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.106138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.106152 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.106169 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.106181 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:40Z","lastTransitionTime":"2025-10-08T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.209206 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.209277 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.209292 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.209313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.209327 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:40Z","lastTransitionTime":"2025-10-08T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.312720 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.312810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.312837 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.312919 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.312946 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:40Z","lastTransitionTime":"2025-10-08T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.416100 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.416264 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.416329 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.416684 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.416737 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:40Z","lastTransitionTime":"2025-10-08T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.452758 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:40 crc kubenswrapper[4744]: E1008 09:12:40.452934 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.520152 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.520259 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.520280 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.520746 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.520981 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:40Z","lastTransitionTime":"2025-10-08T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.624135 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.624201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.624213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.624235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.624251 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:40Z","lastTransitionTime":"2025-10-08T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.726510 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.726549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.726558 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.726572 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.726585 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:40Z","lastTransitionTime":"2025-10-08T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.829902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.829988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.830008 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.830035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.830055 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:40Z","lastTransitionTime":"2025-10-08T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.933353 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.933444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.933462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.933495 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:40 crc kubenswrapper[4744]: I1008 09:12:40.933519 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:40Z","lastTransitionTime":"2025-10-08T09:12:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.038522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.038715 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.038735 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.038898 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.038921 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.142686 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.142786 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.142805 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.142833 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.142853 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.246000 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.246339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.246476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.246585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.246683 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.349873 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.349926 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.349943 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.349964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.349981 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.452245 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.452275 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:41 crc kubenswrapper[4744]: E1008 09:12:41.452472 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:41 crc kubenswrapper[4744]: E1008 09:12:41.452785 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.453330 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:41 crc kubenswrapper[4744]: E1008 09:12:41.453480 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.453868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.454036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.454157 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.454300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.454453 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.557229 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.557280 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.557312 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.557330 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.557340 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.660399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.660463 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.660476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.660497 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.660511 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.763760 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.763814 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.763827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.763847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.763865 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.845065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.845126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.845147 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.845169 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.845183 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: E1008 09:12:41.865021 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:41Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.869176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.869213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.869223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.869242 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.869252 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: E1008 09:12:41.886625 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:41Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.891449 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.891500 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.891512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.891530 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.891542 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: E1008 09:12:41.903904 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:41Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.908550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.908597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.908625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.908646 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.908663 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: E1008 09:12:41.923466 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:41Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.928719 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.928771 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.928781 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.928797 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.928809 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:41 crc kubenswrapper[4744]: E1008 09:12:41.943337 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:41Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:41 crc kubenswrapper[4744]: E1008 09:12:41.943487 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.945823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.945904 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.945927 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.945956 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:41 crc kubenswrapper[4744]: I1008 09:12:41.945976 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:41Z","lastTransitionTime":"2025-10-08T09:12:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.049934 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.049975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.049989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.050015 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.050041 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:42Z","lastTransitionTime":"2025-10-08T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.152888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.152938 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.152951 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.152967 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.152981 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:42Z","lastTransitionTime":"2025-10-08T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.255930 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.256002 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.256019 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.256042 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.256056 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:42Z","lastTransitionTime":"2025-10-08T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.359622 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.359666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.359678 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.359699 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.359711 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:42Z","lastTransitionTime":"2025-10-08T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.452535 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:42 crc kubenswrapper[4744]: E1008 09:12:42.452712 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.463477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.463733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.463755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.463779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.463801 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:42Z","lastTransitionTime":"2025-10-08T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.566530 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.566578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.566589 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.566610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.566624 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:42Z","lastTransitionTime":"2025-10-08T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.669830 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.669895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.669917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.669943 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.669964 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:42Z","lastTransitionTime":"2025-10-08T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.772171 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.772248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.772262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.772280 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.772297 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:42Z","lastTransitionTime":"2025-10-08T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.876180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.876243 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.876257 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.876276 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.876292 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:42Z","lastTransitionTime":"2025-10-08T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.978574 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.978635 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.978648 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.978670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:42 crc kubenswrapper[4744]: I1008 09:12:42.978684 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:42Z","lastTransitionTime":"2025-10-08T09:12:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.082538 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.082629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.082679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.082706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.082723 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:43Z","lastTransitionTime":"2025-10-08T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.185812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.186148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.186557 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.186823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.186997 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:43Z","lastTransitionTime":"2025-10-08T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.289935 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.290300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.290399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.290472 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.290531 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:43Z","lastTransitionTime":"2025-10-08T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.393310 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.393586 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.393657 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.393765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.393870 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:43Z","lastTransitionTime":"2025-10-08T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.453047 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:43 crc kubenswrapper[4744]: E1008 09:12:43.453485 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.453580 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.454057 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:43 crc kubenswrapper[4744]: E1008 09:12:43.454244 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:43 crc kubenswrapper[4744]: E1008 09:12:43.454625 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.497645 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.497712 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.497733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.497757 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.497776 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:43Z","lastTransitionTime":"2025-10-08T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.601578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.601640 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.601664 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.601697 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.601721 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:43Z","lastTransitionTime":"2025-10-08T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.707666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.707775 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.707808 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.707847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.707890 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:43Z","lastTransitionTime":"2025-10-08T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.811815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.811880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.811893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.811910 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.811921 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:43Z","lastTransitionTime":"2025-10-08T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.914918 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.914972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.914987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.915006 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:43 crc kubenswrapper[4744]: I1008 09:12:43.915019 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:43Z","lastTransitionTime":"2025-10-08T09:12:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.018095 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.018152 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.018166 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.018183 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.018194 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:44Z","lastTransitionTime":"2025-10-08T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.121112 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.121172 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.121188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.121212 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.121230 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:44Z","lastTransitionTime":"2025-10-08T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.223794 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.223848 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.223862 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.223881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.223895 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:44Z","lastTransitionTime":"2025-10-08T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.326697 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.327131 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.327352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.327642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.327873 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:44Z","lastTransitionTime":"2025-10-08T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.431069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.431136 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.431154 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.431178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.431199 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:44Z","lastTransitionTime":"2025-10-08T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.452357 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:44 crc kubenswrapper[4744]: E1008 09:12:44.452556 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.453927 4744 scope.go:117] "RemoveContainer" containerID="8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0" Oct 08 09:12:44 crc kubenswrapper[4744]: E1008 09:12:44.454220 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.534166 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.534234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.534253 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.534273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.534289 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:44Z","lastTransitionTime":"2025-10-08T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.637562 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.637617 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.637630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.637649 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.637664 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:44Z","lastTransitionTime":"2025-10-08T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.740882 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.740938 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.740952 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.740972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.740985 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:44Z","lastTransitionTime":"2025-10-08T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.843737 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.843781 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.843793 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.843810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.843823 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:44Z","lastTransitionTime":"2025-10-08T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.946663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.946731 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.946754 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.946778 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:44 crc kubenswrapper[4744]: I1008 09:12:44.946800 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:44Z","lastTransitionTime":"2025-10-08T09:12:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.049784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.049871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.049884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.049902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.049914 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:45Z","lastTransitionTime":"2025-10-08T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.152684 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.152766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.152787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.152812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.152834 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:45Z","lastTransitionTime":"2025-10-08T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.255149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.255223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.255239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.255262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.255276 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:45Z","lastTransitionTime":"2025-10-08T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.358194 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.358283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.358303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.358334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.358425 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:45Z","lastTransitionTime":"2025-10-08T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.453009 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:45 crc kubenswrapper[4744]: E1008 09:12:45.453679 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.454174 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.454354 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:45 crc kubenswrapper[4744]: E1008 09:12:45.455306 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:45 crc kubenswrapper[4744]: E1008 09:12:45.455554 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.462458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.462622 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.462652 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.462736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.462810 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:45Z","lastTransitionTime":"2025-10-08T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.475341 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.493652 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99851b67-455c-484f-9b91-fb10109d9aab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f870e3d341312ab4d6a6faad06746ae0fbac51d385a11767119039cb5440ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e979533e4efeb3bf49255888b07c7aaf770c00a2fb105991b9ace6ce5422d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49105d17503dca06928dea1ef7758199019bd0dea2af285149f0d81f06d6fba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.512075 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.527340 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.540974 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.554774 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.565812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.565851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.565899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.565917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.565931 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:45Z","lastTransitionTime":"2025-10-08T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.570590 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.584194 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.596964 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.610280 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.623227 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.636555 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.651023 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.665055 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.669008 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.669038 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.669050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.669067 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.669080 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:45Z","lastTransitionTime":"2025-10-08T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.686507 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:32Z\\\",\\\"message\\\":\\\"o:386] Retry successful for *v1.Pod openshift-multus/multus-dldlq after 0 failed attempt(s)\\\\nI1008 09:12:32.351708 6275 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 09:12:32.351717 6275 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j\\\\nI1008 09:12:32.351729 6275 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 09:12:32.351734 6275 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j in node crc\\\\nI1008 09:12:32.351704 6275 event.go:377] Event(v1.ObjectReference{Kind:\\\\\\\"Pod\\\\\\\", Namespace:\\\\\\\"openshift-multus\\\\\\\", Name:\\\\\\\"network-metrics-daemon-cdcbz\\\\\\\", UID:\\\\\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26893\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.697987 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.708492 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:45Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.771703 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.771766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.771777 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.771793 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.771804 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:45Z","lastTransitionTime":"2025-10-08T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.875410 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.875456 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.875467 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.875483 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.875494 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:45Z","lastTransitionTime":"2025-10-08T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.978522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.978572 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.978588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.978608 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:45 crc kubenswrapper[4744]: I1008 09:12:45.978624 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:45Z","lastTransitionTime":"2025-10-08T09:12:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.082476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.082534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.082549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.082571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.082587 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:46Z","lastTransitionTime":"2025-10-08T09:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.185478 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.185539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.185565 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.185594 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.185617 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:46Z","lastTransitionTime":"2025-10-08T09:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.289030 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.289094 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.289113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.289144 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.289162 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:46Z","lastTransitionTime":"2025-10-08T09:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.391673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.391710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.391719 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.391735 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.391744 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:46Z","lastTransitionTime":"2025-10-08T09:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.453055 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:46 crc kubenswrapper[4744]: E1008 09:12:46.453262 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.494158 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.494206 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.494216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.494231 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.494242 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:46Z","lastTransitionTime":"2025-10-08T09:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.597409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.597483 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.597493 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.597507 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.597516 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:46Z","lastTransitionTime":"2025-10-08T09:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.701824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.701897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.701923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.701953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.701994 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:46Z","lastTransitionTime":"2025-10-08T09:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.805543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.805604 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.805616 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.805636 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.805646 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:46Z","lastTransitionTime":"2025-10-08T09:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.908678 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.908774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.908804 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.908846 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:46 crc kubenswrapper[4744]: I1008 09:12:46.908875 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:46Z","lastTransitionTime":"2025-10-08T09:12:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.012254 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.012741 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.013079 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.013334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.013636 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:47Z","lastTransitionTime":"2025-10-08T09:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.116811 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.116865 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.116885 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.116910 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.116927 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:47Z","lastTransitionTime":"2025-10-08T09:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.220172 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.220235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.220254 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.220279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.220298 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:47Z","lastTransitionTime":"2025-10-08T09:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.322455 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.322499 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.322510 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.322527 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.322539 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:47Z","lastTransitionTime":"2025-10-08T09:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.425837 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.425895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.425913 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.425936 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.425960 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:47Z","lastTransitionTime":"2025-10-08T09:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.452531 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.452595 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.452530 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:47 crc kubenswrapper[4744]: E1008 09:12:47.452743 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:47 crc kubenswrapper[4744]: E1008 09:12:47.452867 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:47 crc kubenswrapper[4744]: E1008 09:12:47.452945 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.528996 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.529028 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.529039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.529066 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.529075 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:47Z","lastTransitionTime":"2025-10-08T09:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.631788 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.631845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.631867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.631894 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.631918 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:47Z","lastTransitionTime":"2025-10-08T09:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.734724 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.734781 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.734805 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.734836 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.734862 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:47Z","lastTransitionTime":"2025-10-08T09:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.838173 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.838243 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.838263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.838291 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.838316 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:47Z","lastTransitionTime":"2025-10-08T09:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.941076 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.941117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.941127 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.941143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:47 crc kubenswrapper[4744]: I1008 09:12:47.941156 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:47Z","lastTransitionTime":"2025-10-08T09:12:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.044516 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.044566 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.044578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.044596 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.044611 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:48Z","lastTransitionTime":"2025-10-08T09:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.146850 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.146992 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.147010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.147036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.147054 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:48Z","lastTransitionTime":"2025-10-08T09:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.250629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.250692 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.250707 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.250727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.250740 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:48Z","lastTransitionTime":"2025-10-08T09:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.353880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.353929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.353941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.353958 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.353972 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:48Z","lastTransitionTime":"2025-10-08T09:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.452654 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:48 crc kubenswrapper[4744]: E1008 09:12:48.452862 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.457283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.457347 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.457410 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.457445 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.457464 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:48Z","lastTransitionTime":"2025-10-08T09:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.560393 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.560465 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.560491 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.560525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.560548 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:48Z","lastTransitionTime":"2025-10-08T09:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.663162 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.663206 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.663215 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.663228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.663237 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:48Z","lastTransitionTime":"2025-10-08T09:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.766897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.766968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.766986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.767011 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.767030 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:48Z","lastTransitionTime":"2025-10-08T09:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.870517 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.870569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.870580 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.870600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.870616 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:48Z","lastTransitionTime":"2025-10-08T09:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.973441 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.973492 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.973507 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.973529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:48 crc kubenswrapper[4744]: I1008 09:12:48.973543 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:48Z","lastTransitionTime":"2025-10-08T09:12:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.076077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.076113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.076123 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.076138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.076150 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:49Z","lastTransitionTime":"2025-10-08T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.179295 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.179465 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.179485 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.179511 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.179530 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:49Z","lastTransitionTime":"2025-10-08T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.287704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.287873 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.287903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.288040 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.288061 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:49Z","lastTransitionTime":"2025-10-08T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.391389 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.391437 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.391447 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.391466 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.391479 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:49Z","lastTransitionTime":"2025-10-08T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.453113 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.453164 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.453197 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:49 crc kubenswrapper[4744]: E1008 09:12:49.453270 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:49 crc kubenswrapper[4744]: E1008 09:12:49.453474 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:49 crc kubenswrapper[4744]: E1008 09:12:49.453665 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.494014 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.494056 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.494068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.494082 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.494093 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:49Z","lastTransitionTime":"2025-10-08T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.596760 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.596792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.596801 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.596815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.596825 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:49Z","lastTransitionTime":"2025-10-08T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.699870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.699905 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.699917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.699936 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.699951 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:49Z","lastTransitionTime":"2025-10-08T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.802634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.802693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.802702 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.802730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.802740 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:49Z","lastTransitionTime":"2025-10-08T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.905326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.905382 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.905397 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.905414 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:49 crc kubenswrapper[4744]: I1008 09:12:49.905427 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:49Z","lastTransitionTime":"2025-10-08T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.007925 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.007973 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.007986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.008007 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.008023 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:50Z","lastTransitionTime":"2025-10-08T09:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.110949 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.110990 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.111004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.111021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.111035 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:50Z","lastTransitionTime":"2025-10-08T09:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.214096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.214138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.214153 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.214171 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.214183 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:50Z","lastTransitionTime":"2025-10-08T09:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.317106 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.317187 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.317203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.317228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.317244 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:50Z","lastTransitionTime":"2025-10-08T09:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.419797 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.419856 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.419873 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.419898 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.419917 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:50Z","lastTransitionTime":"2025-10-08T09:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.452292 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:50 crc kubenswrapper[4744]: E1008 09:12:50.452468 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.522191 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.522236 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.522250 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.522268 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.522280 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:50Z","lastTransitionTime":"2025-10-08T09:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.625483 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.625538 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.625552 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.625572 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.625586 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:50Z","lastTransitionTime":"2025-10-08T09:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.728830 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.728877 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.728891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.728911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.728923 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:50Z","lastTransitionTime":"2025-10-08T09:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.832508 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.832556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.832567 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.832583 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.832596 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:50Z","lastTransitionTime":"2025-10-08T09:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.934728 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.934807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.934834 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.934868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:50 crc kubenswrapper[4744]: I1008 09:12:50.935471 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:50Z","lastTransitionTime":"2025-10-08T09:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.039027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.039088 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.039111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.039137 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.039156 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:51Z","lastTransitionTime":"2025-10-08T09:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.142874 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.142939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.142958 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.142986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.143009 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:51Z","lastTransitionTime":"2025-10-08T09:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.245816 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.245881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.245949 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.245974 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.245988 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:51Z","lastTransitionTime":"2025-10-08T09:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.349477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.349537 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.349554 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.349573 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.349587 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:51Z","lastTransitionTime":"2025-10-08T09:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.452156 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.452259 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:51 crc kubenswrapper[4744]: E1008 09:12:51.452298 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:51 crc kubenswrapper[4744]: E1008 09:12:51.452469 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.452657 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.452718 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.452731 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.452749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.452761 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:51Z","lastTransitionTime":"2025-10-08T09:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.452176 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:51 crc kubenswrapper[4744]: E1008 09:12:51.452894 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.556582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.556638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.556654 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.556675 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.556691 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:51Z","lastTransitionTime":"2025-10-08T09:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.659728 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.659800 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.659823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.659851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.659871 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:51Z","lastTransitionTime":"2025-10-08T09:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.676518 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:51 crc kubenswrapper[4744]: E1008 09:12:51.676772 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:51 crc kubenswrapper[4744]: E1008 09:12:51.676885 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs podName:be364129-0f0b-41d4-b5f8-1b1def9a1c38 nodeName:}" failed. No retries permitted until 2025-10-08 09:13:23.676858055 +0000 UTC m=+98.924503484 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs") pod "network-metrics-daemon-cdcbz" (UID: "be364129-0f0b-41d4-b5f8-1b1def9a1c38") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.762993 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.763072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.763088 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.763131 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.763146 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:51Z","lastTransitionTime":"2025-10-08T09:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.866340 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.866461 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.866487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.866518 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.866539 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:51Z","lastTransitionTime":"2025-10-08T09:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.946080 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.946455 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.946576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.946680 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.946771 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:51Z","lastTransitionTime":"2025-10-08T09:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:51 crc kubenswrapper[4744]: E1008 09:12:51.963035 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:51Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.967034 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.967363 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.967492 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.967597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.967682 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:51Z","lastTransitionTime":"2025-10-08T09:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:51 crc kubenswrapper[4744]: E1008 09:12:51.982343 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:51Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.985833 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.985881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.985895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.985914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:51 crc kubenswrapper[4744]: I1008 09:12:51.985926 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:51Z","lastTransitionTime":"2025-10-08T09:12:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:52 crc kubenswrapper[4744]: E1008 09:12:52.006163 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:52Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.011908 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.011969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.011981 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.012000 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.012014 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:52Z","lastTransitionTime":"2025-10-08T09:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:52 crc kubenswrapper[4744]: E1008 09:12:52.025564 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:52Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.029442 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.029473 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.029483 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.029514 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.029524 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:52Z","lastTransitionTime":"2025-10-08T09:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:52 crc kubenswrapper[4744]: E1008 09:12:52.045312 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:52Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:52 crc kubenswrapper[4744]: E1008 09:12:52.045526 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.049424 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.049492 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.049509 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.049556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.049575 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:52Z","lastTransitionTime":"2025-10-08T09:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.153518 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.153578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.153599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.153625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.153643 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:52Z","lastTransitionTime":"2025-10-08T09:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.256799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.256832 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.256842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.256855 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.256866 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:52Z","lastTransitionTime":"2025-10-08T09:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.359550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.359586 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.359595 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.359611 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.359621 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:52Z","lastTransitionTime":"2025-10-08T09:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.452760 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:52 crc kubenswrapper[4744]: E1008 09:12:52.452920 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.463155 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.463186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.463196 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.463211 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.463224 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:52Z","lastTransitionTime":"2025-10-08T09:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.566411 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.566458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.566467 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.566487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.566498 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:52Z","lastTransitionTime":"2025-10-08T09:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.670066 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.670138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.670158 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.670181 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.670199 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:52Z","lastTransitionTime":"2025-10-08T09:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.773354 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.773431 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.773449 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.773472 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.773489 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:52Z","lastTransitionTime":"2025-10-08T09:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.877879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.877990 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.878020 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.878054 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.878081 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:52Z","lastTransitionTime":"2025-10-08T09:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.981669 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.981725 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.981743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.981769 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:52 crc kubenswrapper[4744]: I1008 09:12:52.981786 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:52Z","lastTransitionTime":"2025-10-08T09:12:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.085414 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.085469 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.085484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.085509 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.085523 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:53Z","lastTransitionTime":"2025-10-08T09:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.188168 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.188214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.188224 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.188241 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.188252 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:53Z","lastTransitionTime":"2025-10-08T09:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.291286 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.291342 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.291360 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.291403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.291418 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:53Z","lastTransitionTime":"2025-10-08T09:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.394299 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.394391 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.394404 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.394439 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.394449 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:53Z","lastTransitionTime":"2025-10-08T09:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.453138 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.453142 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.453414 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:53 crc kubenswrapper[4744]: E1008 09:12:53.453501 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:53 crc kubenswrapper[4744]: E1008 09:12:53.453333 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:53 crc kubenswrapper[4744]: E1008 09:12:53.453636 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.497276 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.497316 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.497328 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.497346 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.497357 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:53Z","lastTransitionTime":"2025-10-08T09:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.603864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.603906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.603917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.603936 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.603950 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:53Z","lastTransitionTime":"2025-10-08T09:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.707283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.707339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.707350 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.707382 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.707393 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:53Z","lastTransitionTime":"2025-10-08T09:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.810486 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.810532 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.810555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.810576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.810590 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:53Z","lastTransitionTime":"2025-10-08T09:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.914312 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.914405 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.914424 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.914450 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.914469 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:53Z","lastTransitionTime":"2025-10-08T09:12:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.952807 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dldlq_7ea48507-1f33-46fc-ab34-c6e083dd86a9/kube-multus/0.log" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.952892 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ea48507-1f33-46fc-ab34-c6e083dd86a9" containerID="706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f" exitCode=1 Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.952952 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dldlq" event={"ID":"7ea48507-1f33-46fc-ab34-c6e083dd86a9","Type":"ContainerDied","Data":"706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f"} Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.953633 4744 scope.go:117] "RemoveContainer" containerID="706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.973100 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:53Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:53 crc kubenswrapper[4744]: I1008 09:12:53.991148 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:53Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.005256 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.017409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.017446 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.017458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.017473 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.017487 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:54Z","lastTransitionTime":"2025-10-08T09:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.026058 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.048349 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99851b67-455c-484f-9b91-fb10109d9aab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f870e3d341312ab4d6a6faad06746ae0fbac51d385a11767119039cb5440ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e979533e4efeb3bf49255888b07c7aaf770c00a2fb105991b9ace6ce5422d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49105d17503dca06928dea1ef7758199019bd0dea2af285149f0d81f06d6fba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.064318 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.080754 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.096188 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.105498 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.116133 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.120469 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.120548 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.120561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.120582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.120596 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:54Z","lastTransitionTime":"2025-10-08T09:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.127730 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:53Z\\\",\\\"message\\\":\\\"2025-10-08T09:12:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6\\\\n2025-10-08T09:12:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6 to /host/opt/cni/bin/\\\\n2025-10-08T09:12:08Z [verbose] multus-daemon started\\\\n2025-10-08T09:12:08Z [verbose] Readiness Indicator file check\\\\n2025-10-08T09:12:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.147570 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.166222 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.184952 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.198283 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.217964 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:32Z\\\",\\\"message\\\":\\\"o:386] Retry successful for *v1.Pod openshift-multus/multus-dldlq after 0 failed attempt(s)\\\\nI1008 09:12:32.351708 6275 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 09:12:32.351717 6275 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j\\\\nI1008 09:12:32.351729 6275 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 09:12:32.351734 6275 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j in node crc\\\\nI1008 09:12:32.351704 6275 event.go:377] Event(v1.ObjectReference{Kind:\\\\\\\"Pod\\\\\\\", Namespace:\\\\\\\"openshift-multus\\\\\\\", Name:\\\\\\\"network-metrics-daemon-cdcbz\\\\\\\", UID:\\\\\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26893\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.222497 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.222537 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.222547 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.222563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.222573 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:54Z","lastTransitionTime":"2025-10-08T09:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.233955 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.325623 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.325665 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.325677 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.325693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.325703 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:54Z","lastTransitionTime":"2025-10-08T09:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.429121 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.429322 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.429448 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.429551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.429637 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:54Z","lastTransitionTime":"2025-10-08T09:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.452388 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:54 crc kubenswrapper[4744]: E1008 09:12:54.452569 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.533507 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.534163 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.534346 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.534558 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.534717 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:54Z","lastTransitionTime":"2025-10-08T09:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.637976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.638039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.638050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.638069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.638085 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:54Z","lastTransitionTime":"2025-10-08T09:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.741627 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.741679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.741695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.741717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.741733 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:54Z","lastTransitionTime":"2025-10-08T09:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.844194 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.844243 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.844255 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.844270 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.844289 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:54Z","lastTransitionTime":"2025-10-08T09:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.946975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.947066 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.947081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.947103 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.947120 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:54Z","lastTransitionTime":"2025-10-08T09:12:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.957433 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dldlq_7ea48507-1f33-46fc-ab34-c6e083dd86a9/kube-multus/0.log" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.957500 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dldlq" event={"ID":"7ea48507-1f33-46fc-ab34-c6e083dd86a9","Type":"ContainerStarted","Data":"6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75"} Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.972308 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:54 crc kubenswrapper[4744]: I1008 09:12:54.987429 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.001176 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:54Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.014349 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.026823 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.037955 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.050724 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.050760 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.050772 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.050788 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.050800 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:55Z","lastTransitionTime":"2025-10-08T09:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.056950 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:53Z\\\",\\\"message\\\":\\\"2025-10-08T09:12:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6\\\\n2025-10-08T09:12:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6 to /host/opt/cni/bin/\\\\n2025-10-08T09:12:08Z [verbose] multus-daemon started\\\\n2025-10-08T09:12:08Z [verbose] Readiness Indicator file check\\\\n2025-10-08T09:12:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.070802 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.087178 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.101013 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.114749 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.136578 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:32Z\\\",\\\"message\\\":\\\"o:386] Retry successful for *v1.Pod openshift-multus/multus-dldlq after 0 failed attempt(s)\\\\nI1008 09:12:32.351708 6275 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 09:12:32.351717 6275 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j\\\\nI1008 09:12:32.351729 6275 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 09:12:32.351734 6275 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j in node crc\\\\nI1008 09:12:32.351704 6275 event.go:377] Event(v1.ObjectReference{Kind:\\\\\\\"Pod\\\\\\\", Namespace:\\\\\\\"openshift-multus\\\\\\\", Name:\\\\\\\"network-metrics-daemon-cdcbz\\\\\\\", UID:\\\\\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26893\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.151066 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99851b67-455c-484f-9b91-fb10109d9aab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f870e3d341312ab4d6a6faad06746ae0fbac51d385a11767119039cb5440ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e979533e4efeb3bf49255888b07c7aaf770c00a2fb105991b9ace6ce5422d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49105d17503dca06928dea1ef7758199019bd0dea2af285149f0d81f06d6fba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.153309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.153452 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.153538 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.153632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.153711 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:55Z","lastTransitionTime":"2025-10-08T09:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.168615 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.180387 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.190852 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.207815 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.263605 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.263660 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.263670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.263686 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.263699 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:55Z","lastTransitionTime":"2025-10-08T09:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.366039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.366358 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.366468 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.366564 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.366648 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:55Z","lastTransitionTime":"2025-10-08T09:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.452152 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:55 crc kubenswrapper[4744]: E1008 09:12:55.452307 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.452169 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:55 crc kubenswrapper[4744]: E1008 09:12:55.452391 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.452152 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:55 crc kubenswrapper[4744]: E1008 09:12:55.452468 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.466969 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.469237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.469263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.469273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.469288 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.469298 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:55Z","lastTransitionTime":"2025-10-08T09:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.481089 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.500820 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:32Z\\\",\\\"message\\\":\\\"o:386] Retry successful for *v1.Pod openshift-multus/multus-dldlq after 0 failed attempt(s)\\\\nI1008 09:12:32.351708 6275 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 09:12:32.351717 6275 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j\\\\nI1008 09:12:32.351729 6275 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 09:12:32.351734 6275 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j in node crc\\\\nI1008 09:12:32.351704 6275 event.go:377] Event(v1.ObjectReference{Kind:\\\\\\\"Pod\\\\\\\", Namespace:\\\\\\\"openshift-multus\\\\\\\", Name:\\\\\\\"network-metrics-daemon-cdcbz\\\\\\\", UID:\\\\\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26893\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.517058 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.530540 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.545702 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.560438 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.572065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.572099 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.572111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.572128 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.572141 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:55Z","lastTransitionTime":"2025-10-08T09:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.574941 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99851b67-455c-484f-9b91-fb10109d9aab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f870e3d341312ab4d6a6faad06746ae0fbac51d385a11767119039cb5440ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e979533e4efeb3bf49255888b07c7aaf770c00a2fb105991b9ace6ce5422d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49105d17503dca06928dea1ef7758199019bd0dea2af285149f0d81f06d6fba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.590479 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.605926 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.620436 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.638022 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.651785 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.663973 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:53Z\\\",\\\"message\\\":\\\"2025-10-08T09:12:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6\\\\n2025-10-08T09:12:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6 to /host/opt/cni/bin/\\\\n2025-10-08T09:12:08Z [verbose] multus-daemon started\\\\n2025-10-08T09:12:08Z [verbose] Readiness Indicator file check\\\\n2025-10-08T09:12:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.675514 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.675554 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.675567 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.675588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.675602 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:55Z","lastTransitionTime":"2025-10-08T09:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.678797 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.694618 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.707972 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:12:55Z is after 2025-08-24T17:21:41Z" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.779419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.779484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.779496 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.779514 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.779531 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:55Z","lastTransitionTime":"2025-10-08T09:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.883027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.883128 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.883147 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.883172 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.883191 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:55Z","lastTransitionTime":"2025-10-08T09:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.985501 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.985541 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.985551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.985567 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:55 crc kubenswrapper[4744]: I1008 09:12:55.985578 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:55Z","lastTransitionTime":"2025-10-08T09:12:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.089355 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.089405 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.089415 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.089431 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.089441 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:56Z","lastTransitionTime":"2025-10-08T09:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.192766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.192855 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.192888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.192922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.192949 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:56Z","lastTransitionTime":"2025-10-08T09:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.297667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.297751 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.297765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.297782 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.297801 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:56Z","lastTransitionTime":"2025-10-08T09:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.406686 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.406736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.406754 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.406773 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.406786 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:56Z","lastTransitionTime":"2025-10-08T09:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.452747 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:56 crc kubenswrapper[4744]: E1008 09:12:56.452936 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.510453 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.510501 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.510511 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.510528 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.510540 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:56Z","lastTransitionTime":"2025-10-08T09:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.612928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.612975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.612985 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.613003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.613017 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:56Z","lastTransitionTime":"2025-10-08T09:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.716708 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.716782 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.716805 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.716837 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.716859 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:56Z","lastTransitionTime":"2025-10-08T09:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.820492 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.820531 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.820543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.820560 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.820572 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:56Z","lastTransitionTime":"2025-10-08T09:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.923786 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.923842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.923859 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.923884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:56 crc kubenswrapper[4744]: I1008 09:12:56.923902 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:56Z","lastTransitionTime":"2025-10-08T09:12:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.027835 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.027897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.027916 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.027943 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.027963 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:57Z","lastTransitionTime":"2025-10-08T09:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.130579 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.130643 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.130663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.130690 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.130709 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:57Z","lastTransitionTime":"2025-10-08T09:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.233519 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.233585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.233608 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.233650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.233669 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:57Z","lastTransitionTime":"2025-10-08T09:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.335838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.335869 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.335879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.335892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.335903 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:57Z","lastTransitionTime":"2025-10-08T09:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.445706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.445752 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.445763 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.445779 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.445791 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:57Z","lastTransitionTime":"2025-10-08T09:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.452998 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.453018 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.453124 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:57 crc kubenswrapper[4744]: E1008 09:12:57.453215 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:57 crc kubenswrapper[4744]: E1008 09:12:57.453406 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:57 crc kubenswrapper[4744]: E1008 09:12:57.453515 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.548225 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.548272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.548285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.548303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.548317 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:57Z","lastTransitionTime":"2025-10-08T09:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.650807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.650881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.650895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.650921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.650933 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:57Z","lastTransitionTime":"2025-10-08T09:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.753833 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.753882 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.753900 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.753922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.753934 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:57Z","lastTransitionTime":"2025-10-08T09:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.860326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.860388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.860404 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.860425 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.860438 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:57Z","lastTransitionTime":"2025-10-08T09:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.962991 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.963034 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.963046 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.963091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:57 crc kubenswrapper[4744]: I1008 09:12:57.963103 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:57Z","lastTransitionTime":"2025-10-08T09:12:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.065000 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.065049 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.065063 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.065085 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.065098 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:58Z","lastTransitionTime":"2025-10-08T09:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.167398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.167459 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.167473 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.167496 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.167508 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:58Z","lastTransitionTime":"2025-10-08T09:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.270656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.270711 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.270725 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.270746 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.270759 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:58Z","lastTransitionTime":"2025-10-08T09:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.382422 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.382478 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.382488 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.382504 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.382516 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:58Z","lastTransitionTime":"2025-10-08T09:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.452514 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:12:58 crc kubenswrapper[4744]: E1008 09:12:58.452689 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.485315 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.485351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.485364 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.485409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.485420 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:58Z","lastTransitionTime":"2025-10-08T09:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.589210 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.589272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.589296 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.589329 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.589351 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:58Z","lastTransitionTime":"2025-10-08T09:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.692180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.692249 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.692265 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.692289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.692306 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:58Z","lastTransitionTime":"2025-10-08T09:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.795109 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.795157 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.795168 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.795189 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.795205 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:58Z","lastTransitionTime":"2025-10-08T09:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.898739 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.898858 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.898870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.898893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:58 crc kubenswrapper[4744]: I1008 09:12:58.898906 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:58Z","lastTransitionTime":"2025-10-08T09:12:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.010093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.010664 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.010765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.011112 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.011228 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:59Z","lastTransitionTime":"2025-10-08T09:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.115213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.115272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.115284 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.115305 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.115317 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:59Z","lastTransitionTime":"2025-10-08T09:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.218306 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.218941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.219040 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.219171 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.219255 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:59Z","lastTransitionTime":"2025-10-08T09:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.322677 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.322745 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.322756 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.322776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.322787 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:59Z","lastTransitionTime":"2025-10-08T09:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.425897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.426412 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.426543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.426674 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.426799 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:59Z","lastTransitionTime":"2025-10-08T09:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.452582 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:12:59 crc kubenswrapper[4744]: E1008 09:12:59.452722 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.452756 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.452793 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:12:59 crc kubenswrapper[4744]: E1008 09:12:59.452866 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.453398 4744 scope.go:117] "RemoveContainer" containerID="8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0" Oct 08 09:12:59 crc kubenswrapper[4744]: E1008 09:12:59.453738 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.530136 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.530403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.530475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.530551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.530619 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:59Z","lastTransitionTime":"2025-10-08T09:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.633960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.634002 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.634015 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.634034 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.634045 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:59Z","lastTransitionTime":"2025-10-08T09:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.737341 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.737419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.737436 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.737463 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.737479 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:59Z","lastTransitionTime":"2025-10-08T09:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.840983 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.841024 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.841035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.841054 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.841068 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:59Z","lastTransitionTime":"2025-10-08T09:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.944699 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.944771 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.944784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.944809 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.944826 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:12:59Z","lastTransitionTime":"2025-10-08T09:12:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.980770 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/2.log" Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.984789 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357"} Oct 08 09:12:59 crc kubenswrapper[4744]: I1008 09:12:59.985501 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.030153 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:32Z\\\",\\\"message\\\":\\\"o:386] Retry successful for *v1.Pod openshift-multus/multus-dldlq after 0 failed attempt(s)\\\\nI1008 09:12:32.351708 6275 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 09:12:32.351717 6275 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j\\\\nI1008 09:12:32.351729 6275 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 09:12:32.351734 6275 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j in node crc\\\\nI1008 09:12:32.351704 6275 event.go:377] Event(v1.ObjectReference{Kind:\\\\\\\"Pod\\\\\\\", Namespace:\\\\\\\"openshift-multus\\\\\\\", Name:\\\\\\\"network-metrics-daemon-cdcbz\\\\\\\", UID:\\\\\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26893\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.047896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.047951 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.047966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.047987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.048002 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:00Z","lastTransitionTime":"2025-10-08T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.063644 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.091482 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.106908 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.118767 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99851b67-455c-484f-9b91-fb10109d9aab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f870e3d341312ab4d6a6faad06746ae0fbac51d385a11767119039cb5440ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e979533e4efeb3bf49255888b07c7aaf770c00a2fb105991b9ace6ce5422d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49105d17503dca06928dea1ef7758199019bd0dea2af285149f0d81f06d6fba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.132771 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.147239 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.151205 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.151228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.151237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.151278 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.151290 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:00Z","lastTransitionTime":"2025-10-08T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.162532 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.174340 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.184626 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.197444 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:53Z\\\",\\\"message\\\":\\\"2025-10-08T09:12:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6\\\\n2025-10-08T09:12:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6 to /host/opt/cni/bin/\\\\n2025-10-08T09:12:08Z [verbose] multus-daemon started\\\\n2025-10-08T09:12:08Z [verbose] Readiness Indicator file check\\\\n2025-10-08T09:12:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.209863 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.225551 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.238446 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.254284 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.254332 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.254351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.254434 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.254458 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:00Z","lastTransitionTime":"2025-10-08T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.257350 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.272606 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.293461 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:00Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.358001 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.358052 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.358062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.358081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.358092 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:00Z","lastTransitionTime":"2025-10-08T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.452815 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:00 crc kubenswrapper[4744]: E1008 09:13:00.453044 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.461917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.461979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.461997 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.462028 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.462054 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:00Z","lastTransitionTime":"2025-10-08T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.565041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.565095 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.565107 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.565130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.565142 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:00Z","lastTransitionTime":"2025-10-08T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.668699 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.668757 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.668773 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.668816 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.668830 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:00Z","lastTransitionTime":"2025-10-08T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.772655 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.772716 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.772734 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.772757 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.772774 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:00Z","lastTransitionTime":"2025-10-08T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.876496 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.876578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.876597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.876628 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.876654 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:00Z","lastTransitionTime":"2025-10-08T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.979490 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.979541 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.979743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.979762 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.979775 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:00Z","lastTransitionTime":"2025-10-08T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.990784 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/3.log" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.992552 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/2.log" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.998773 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357" exitCode=1 Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.998842 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357"} Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.998914 4744 scope.go:117] "RemoveContainer" containerID="8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0" Oct 08 09:13:00 crc kubenswrapper[4744]: I1008 09:13:00.999915 4744 scope.go:117] "RemoveContainer" containerID="00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357" Oct 08 09:13:01 crc kubenswrapper[4744]: E1008 09:13:01.000158 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.024742 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.043619 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99851b67-455c-484f-9b91-fb10109d9aab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f870e3d341312ab4d6a6faad06746ae0fbac51d385a11767119039cb5440ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e979533e4efeb3bf49255888b07c7aaf770c00a2fb105991b9ace6ce5422d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49105d17503dca06928dea1ef7758199019bd0dea2af285149f0d81f06d6fba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.062679 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.082578 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.082904 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.082935 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.082948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.082968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.082982 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:01Z","lastTransitionTime":"2025-10-08T09:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.099233 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.114950 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.136570 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:53Z\\\",\\\"message\\\":\\\"2025-10-08T09:12:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6\\\\n2025-10-08T09:12:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6 to /host/opt/cni/bin/\\\\n2025-10-08T09:12:08Z [verbose] multus-daemon started\\\\n2025-10-08T09:12:08Z [verbose] Readiness Indicator file check\\\\n2025-10-08T09:12:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.168571 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.185272 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.185789 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.185965 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.186102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.186218 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.186331 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:01Z","lastTransitionTime":"2025-10-08T09:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.200289 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.215337 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.229893 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.246065 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.270681 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.289538 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.289890 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.290024 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.290161 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.290280 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:01Z","lastTransitionTime":"2025-10-08T09:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.297419 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c23b2d79aa4afd382d313113dda0a7d5b921bea03b99e0804a8e2b6953c99d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:32Z\\\",\\\"message\\\":\\\"o:386] Retry successful for *v1.Pod openshift-multus/multus-dldlq after 0 failed attempt(s)\\\\nI1008 09:12:32.351708 6275 services_controller.go:443] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.110\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1008 09:12:32.351717 6275 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j\\\\nI1008 09:12:32.351729 6275 services_controller.go:444] Built service openshift-operator-lifecycle-manager/package-server-manager-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1008 09:12:32.351734 6275 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j in node crc\\\\nI1008 09:12:32.351704 6275 event.go:377] Event(v1.ObjectReference{Kind:\\\\\\\"Pod\\\\\\\", Namespace:\\\\\\\"openshift-multus\\\\\\\", Name:\\\\\\\"network-metrics-daemon-cdcbz\\\\\\\", UID:\\\\\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26893\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddi\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:13:00Z\\\",\\\"message\\\":\\\"/networking-console-plugin-85b44fc459-gdk6g openshift-network-node-identity/network-node-identity-vrzqb]\\\\nI1008 09:13:00.574115 6614 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 09:13:00.574160 6614 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1008 09:13:00.574207 6614 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1008 09:13:00.574248 6614 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1008 09:13:00.574282 6614 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1008 09:13:00.574316 6614 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1008 09:13:00.574421 6614 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 09:13:00.574579 6614 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.310572 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.321835 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:01Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.393667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.394035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.394138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.394238 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.394305 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:01Z","lastTransitionTime":"2025-10-08T09:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.452677 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:01 crc kubenswrapper[4744]: E1008 09:13:01.453022 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.452806 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:01 crc kubenswrapper[4744]: E1008 09:13:01.453222 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.452747 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:01 crc kubenswrapper[4744]: E1008 09:13:01.453568 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.498061 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.498567 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.498671 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.498762 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.498982 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:01Z","lastTransitionTime":"2025-10-08T09:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.602407 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.602462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.602486 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.602512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.602530 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:01Z","lastTransitionTime":"2025-10-08T09:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.706495 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.706572 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.706598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.706632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.706713 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:01Z","lastTransitionTime":"2025-10-08T09:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.810924 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.810991 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.811010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.811040 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.811061 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:01Z","lastTransitionTime":"2025-10-08T09:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.916110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.916184 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.916203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.916236 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:01 crc kubenswrapper[4744]: I1008 09:13:01.916257 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:01Z","lastTransitionTime":"2025-10-08T09:13:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.005950 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/3.log" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.012760 4744 scope.go:117] "RemoveContainer" containerID="00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357" Oct 08 09:13:02 crc kubenswrapper[4744]: E1008 09:13:02.013120 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.018579 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.018642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.018667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.018698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.018724 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.033127 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.053890 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:53Z\\\",\\\"message\\\":\\\"2025-10-08T09:12:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6\\\\n2025-10-08T09:12:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6 to /host/opt/cni/bin/\\\\n2025-10-08T09:12:08Z [verbose] multus-daemon started\\\\n2025-10-08T09:12:08Z [verbose] Readiness Indicator file check\\\\n2025-10-08T09:12:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.071911 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.089295 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.101440 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.101485 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.101506 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.101530 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.101546 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.109536 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: E1008 09:13:02.118260 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.123031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.123079 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.123093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.123117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.123130 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.126529 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: E1008 09:13:02.137982 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.138844 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.142943 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.143000 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.143015 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.143038 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.143055 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.156014 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: E1008 09:13:02.156464 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.161140 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.161177 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.161190 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.161209 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.161220 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.174348 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: E1008 09:13:02.174618 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.179218 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.179452 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.179575 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.179709 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.179842 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: E1008 09:13:02.193710 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T09:13:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c8c6e8b-9d7f-471f-a44b-640abd8b4177\\\",\\\"systemUUID\\\":\\\"f5297ca0-7781-4d45-97fe-e5122f26cce4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: E1008 09:13:02.193887 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.196593 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.196906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.196992 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.197077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.197164 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.198068 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:13:00Z\\\",\\\"message\\\":\\\"/networking-console-plugin-85b44fc459-gdk6g openshift-network-node-identity/network-node-identity-vrzqb]\\\\nI1008 09:13:00.574115 6614 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 09:13:00.574160 6614 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1008 09:13:00.574207 6614 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1008 09:13:00.574248 6614 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1008 09:13:00.574282 6614 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1008 09:13:00.574316 6614 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1008 09:13:00.574421 6614 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 09:13:00.574579 6614 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.212129 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.228850 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.244765 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.255822 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99851b67-455c-484f-9b91-fb10109d9aab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f870e3d341312ab4d6a6faad06746ae0fbac51d385a11767119039cb5440ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e979533e4efeb3bf49255888b07c7aaf770c00a2fb105991b9ace6ce5422d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49105d17503dca06928dea1ef7758199019bd0dea2af285149f0d81f06d6fba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.266815 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.278870 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.291329 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:02Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.299499 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.299561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.299573 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.299599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.299610 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.403462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.403535 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.403553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.403581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.403601 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.453171 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:02 crc kubenswrapper[4744]: E1008 09:13:02.453364 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.507902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.508006 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.508025 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.508059 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.508080 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.611860 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.611931 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.611950 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.611977 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.611996 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.715356 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.715431 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.715442 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.715811 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.715846 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.820201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.820284 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.820304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.820336 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.820362 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.924228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.924304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.924325 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.924353 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:02 crc kubenswrapper[4744]: I1008 09:13:02.924415 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:02Z","lastTransitionTime":"2025-10-08T09:13:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.027054 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.027093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.027105 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.027125 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.027135 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:03Z","lastTransitionTime":"2025-10-08T09:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.130225 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.130289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.130304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.130323 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.130341 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:03Z","lastTransitionTime":"2025-10-08T09:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.233057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.233099 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.233108 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.233125 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.233135 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:03Z","lastTransitionTime":"2025-10-08T09:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.341249 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.341361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.341413 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.341440 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.341463 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:03Z","lastTransitionTime":"2025-10-08T09:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.446267 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.446317 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.446328 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.446345 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.446359 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:03Z","lastTransitionTime":"2025-10-08T09:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.453058 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.453093 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.453181 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:03 crc kubenswrapper[4744]: E1008 09:13:03.453248 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:03 crc kubenswrapper[4744]: E1008 09:13:03.453485 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:03 crc kubenswrapper[4744]: E1008 09:13:03.453740 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.549942 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.550011 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.550031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.550060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.550080 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:03Z","lastTransitionTime":"2025-10-08T09:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.653223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.653324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.653347 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.653413 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.653441 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:03Z","lastTransitionTime":"2025-10-08T09:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.757360 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.758007 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.758025 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.758049 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.758066 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:03Z","lastTransitionTime":"2025-10-08T09:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.861169 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.861215 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.861229 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.861251 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.861265 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:03Z","lastTransitionTime":"2025-10-08T09:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.965284 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.965360 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.965419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.965457 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:03 crc kubenswrapper[4744]: I1008 09:13:03.965481 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:03Z","lastTransitionTime":"2025-10-08T09:13:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.068550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.068632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.068657 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.068691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.068714 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:04Z","lastTransitionTime":"2025-10-08T09:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.172673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.172742 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.172760 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.172796 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.172816 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:04Z","lastTransitionTime":"2025-10-08T09:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.276149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.276233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.276270 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.276303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.276324 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:04Z","lastTransitionTime":"2025-10-08T09:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.382132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.382222 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.382242 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.382271 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.382290 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:04Z","lastTransitionTime":"2025-10-08T09:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.453139 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:04 crc kubenswrapper[4744]: E1008 09:13:04.453315 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.485831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.485881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.485899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.485925 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.485940 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:04Z","lastTransitionTime":"2025-10-08T09:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.588836 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.588904 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.588923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.588954 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.588971 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:04Z","lastTransitionTime":"2025-10-08T09:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.691471 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.691522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.691537 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.691554 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.691564 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:04Z","lastTransitionTime":"2025-10-08T09:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.795767 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.795809 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.795823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.795891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.795909 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:04Z","lastTransitionTime":"2025-10-08T09:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.898708 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.898748 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.898758 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.898776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:04 crc kubenswrapper[4744]: I1008 09:13:04.898786 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:04Z","lastTransitionTime":"2025-10-08T09:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.001259 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.001549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.001561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.001576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.001584 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:05Z","lastTransitionTime":"2025-10-08T09:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.108585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.108620 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.108630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.108644 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.108654 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:05Z","lastTransitionTime":"2025-10-08T09:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.210448 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.210500 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.210512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.210529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.210541 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:05Z","lastTransitionTime":"2025-10-08T09:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.314077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.314121 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.314130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.314145 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.314155 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:05Z","lastTransitionTime":"2025-10-08T09:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.416640 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.416701 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.416711 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.416724 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.416734 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:05Z","lastTransitionTime":"2025-10-08T09:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.452565 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.452589 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:05 crc kubenswrapper[4744]: E1008 09:13:05.452686 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:05 crc kubenswrapper[4744]: E1008 09:13:05.452743 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.452786 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:05 crc kubenswrapper[4744]: E1008 09:13:05.452832 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.472688 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ed33c9d9121afbd72b66a504cae906247141cddb1418fcd178d7ef1753689ffc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7e0527d437c61f7000aaa3d63059a266e88a94c5b9bf7a18b780fc34957ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.494190 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bfffa6a9-a014-46b8-b4de-3261522cc21e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa74b3a987f509f03e747edff49002ba2162005df23806088bc73d667a9bf564\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cd5c4d234ed5abad40c055942108f4b41068e3d8ffba05b26e470e25615bc9f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f904e64e1c64977ed31d749e3c0e1add1a032afe850d55ad5f1113b13153865b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94f66b7bd09a5ad824eea47bfbf47cc59a7076ee9e94b7a5065719a329b82f22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7e857e79b17c3a32b37456a85600f8e7b459c408cdae5e4e5c6c5e40c409b99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb9737dfd4a348d3db74ef44df2a2cccded9f9a6a53d2973df74f3dea405fd46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4493e04c98f304b53aa47daf3295cb03e2dbdbf270f7002d4649b9465c219a08\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ctjzs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-f6z7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.521401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.521443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.521476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.521493 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.521504 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:05Z","lastTransitionTime":"2025-10-08T09:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.521391 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a9bf215-f18b-4732-b12c-0355a3a0a15f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:13:00Z\\\",\\\"message\\\":\\\"/networking-console-plugin-85b44fc459-gdk6g openshift-network-node-identity/network-node-identity-vrzqb]\\\\nI1008 09:13:00.574115 6614 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 09:13:00.574160 6614 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1008 09:13:00.574207 6614 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1008 09:13:00.574248 6614 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI1008 09:13:00.574282 6614 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1008 09:13:00.574316 6614 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1008 09:13:00.574421 6614 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 09:13:00.574579 6614 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q28tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tmmnj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.535623 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0faba02-dfa7-4a4f-9697-05e0321ed28a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b6068e68668c3823a3b8ccde0525e6c06819c86e9b4cddf62bcd02432ecb6ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69cafc2782f80ac2c03a1d415f6956946891df037060ab715636420949e20037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jkf9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jb62j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.548087 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be364129-0f0b-41d4-b5f8-1b1def9a1c38\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9hkbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cdcbz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.562456 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-j2skv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56790493-3c4e-457d-bf41-aa9a7fb680ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b43acfa88f7937348dff7496fca0fb283a714621b272efb6e292ca94e21353a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l689j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-j2skv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.577301 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4866dc16-3b86-4323-95c4-9521e89bf9d8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17ad3a959569e75ba26003c33936c438e75f71d2ee5d349e4d97b101d6d2fc2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45842bbb14a8db293db8a917f9d79e8302a7d9f3d0c3d51283cce4a64b23817f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0eb3ae1c4d7095caa9be57af46184ebb52a5e0ac9ef1993a583ccc9bb57edbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e30fd2181c3095379c3f3b7905ac9e9ed8685347731543fea894a41b27758958\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e30f37d2d7ea510a7b69adc8c5e18b39c39b9dd3397c74da7b9a8ef61575c972\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T09:12:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 09:11:59.098765 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 09:11:59.099753 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3897077811/tls.crt::/tmp/serving-cert-3897077811/tls.key\\\\\\\"\\\\nI1008 09:12:04.637713 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 09:12:04.641844 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 09:12:04.641877 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 09:12:04.641913 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 09:12:04.641924 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 09:12:04.652168 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 09:12:04.652208 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652214 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 09:12:04.652219 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 09:12:04.652224 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 09:12:04.652227 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 09:12:04.652231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 09:12:04.652447 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 09:12:04.656645 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc03c2b0c542c529ca2f18e857545f69c1306c11230cf3fcfd685cd5c40df5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://408130e252044c928903b394e2167624da04788bb54384b877fb77c41aacf340\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.593154 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99851b67-455c-484f-9b91-fb10109d9aab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64f870e3d341312ab4d6a6faad06746ae0fbac51d385a11767119039cb5440ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e979533e4efeb3bf49255888b07c7aaf770c00a2fb105991b9ace6ce5422d5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49105d17503dca06928dea1ef7758199019bd0dea2af285149f0d81f06d6fba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a70d8fabac79a26fbb1e431647ff4ce7606347e303dbf43fefde74bb66af35b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.608797 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.621398 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b42b65c6f015521169d017714b0ff45a7ef7bbfdd93923e1c0f79870099c8c2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.624984 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.625018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.625027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.625041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.625052 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:05Z","lastTransitionTime":"2025-10-08T09:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.635200 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.645554 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-rgsf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305091b5-c89f-49b3-a180-4b304261fc7c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://84699686ee349f6e6e5d38ac23f14fec95dc1a88837a56eb84a965e281a2acd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxjrj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rgsf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.660057 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7da745069c2fa3a72aa5582f8a2221275b42178e77586b80b495b9a26e5fccb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t7mf2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ttxzq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.674975 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dldlq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ea48507-1f33-46fc-ab34-c6e083dd86a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T09:12:53Z\\\",\\\"message\\\":\\\"2025-10-08T09:12:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6\\\\n2025-10-08T09:12:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6c4af5b6-1a39-40d7-a87a-b9b423a067d6 to /host/opt/cni/bin/\\\\n2025-10-08T09:12:08Z [verbose] multus-daemon started\\\\n2025-10-08T09:12:08Z [verbose] Readiness Indicator file check\\\\n2025-10-08T09:12:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6n4nk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:12:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dldlq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.689088 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3052eb00-be1f-4ace-ad1c-e9f6e050d450\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e92636ec67dd107b320b5edad399fab190edde0a677a25e5a40937efe769665c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://07eedaf075d1818263828df846826db9dc6964b4332afe03c4e6b366a3de8b71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57e79e7aea22b7a9359f501372b2bfaf8aa7eb0701a98c631ae4f379a4fa5654\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb980b0f9209f26137f51cea9ff454526ac756235f2ee7fea1dba5ec30f1a7f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T09:11:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.701707 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a576426f1abd3f719a7617e069512ce4f8bbc7b6309da1c27ea4902721b30baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T09:12:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.711837 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T09:12:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T09:13:05Z is after 2025-08-24T17:21:41Z" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.727852 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.727880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.727888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.727903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.727912 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:05Z","lastTransitionTime":"2025-10-08T09:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.830688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.831279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.831481 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.831721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.831873 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:05Z","lastTransitionTime":"2025-10-08T09:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.935404 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.935485 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.935503 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.935526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:05 crc kubenswrapper[4744]: I1008 09:13:05.935545 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:05Z","lastTransitionTime":"2025-10-08T09:13:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.038098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.038160 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.038177 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.038198 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.038216 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:06Z","lastTransitionTime":"2025-10-08T09:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.141450 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.141552 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.141580 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.141616 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.141643 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:06Z","lastTransitionTime":"2025-10-08T09:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.245921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.245971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.245980 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.245998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.246009 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:06Z","lastTransitionTime":"2025-10-08T09:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.350300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.350362 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.350398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.350419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.350433 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:06Z","lastTransitionTime":"2025-10-08T09:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.452728 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:06 crc kubenswrapper[4744]: E1008 09:13:06.452906 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.456192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.456283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.456323 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.456358 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.456416 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:06Z","lastTransitionTime":"2025-10-08T09:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.474582 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.558815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.558884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.558900 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.558924 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.558951 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:06Z","lastTransitionTime":"2025-10-08T09:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.662574 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.662669 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.662694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.662733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.662760 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:06Z","lastTransitionTime":"2025-10-08T09:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.766180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.766263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.766283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.766318 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.766345 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:06Z","lastTransitionTime":"2025-10-08T09:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.870262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.870343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.870413 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.870444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.870465 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:06Z","lastTransitionTime":"2025-10-08T09:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.973885 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.973930 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.973940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.973959 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:06 crc kubenswrapper[4744]: I1008 09:13:06.973970 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:06Z","lastTransitionTime":"2025-10-08T09:13:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.078115 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.078227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.078239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.078262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.078276 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:07Z","lastTransitionTime":"2025-10-08T09:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.182819 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.182901 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.182916 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.182935 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.182948 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:07Z","lastTransitionTime":"2025-10-08T09:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.286049 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.286104 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.286114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.286136 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.286151 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:07Z","lastTransitionTime":"2025-10-08T09:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.388740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.388780 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.388788 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.388801 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.388812 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:07Z","lastTransitionTime":"2025-10-08T09:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.453188 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.453362 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:07 crc kubenswrapper[4744]: E1008 09:13:07.453574 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.453864 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:07 crc kubenswrapper[4744]: E1008 09:13:07.454137 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:07 crc kubenswrapper[4744]: E1008 09:13:07.454306 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.491740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.492046 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.492224 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.492295 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.492356 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:07Z","lastTransitionTime":"2025-10-08T09:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.594164 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.594210 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.594226 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.594247 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.594265 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:07Z","lastTransitionTime":"2025-10-08T09:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.696281 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.696534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.696597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.696660 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.696725 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:07Z","lastTransitionTime":"2025-10-08T09:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.798662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.798703 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.798716 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.798732 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.798747 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:07Z","lastTransitionTime":"2025-10-08T09:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.901465 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.901531 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.901548 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.901574 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:07 crc kubenswrapper[4744]: I1008 09:13:07.901591 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:07Z","lastTransitionTime":"2025-10-08T09:13:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.004177 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.004218 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.004226 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.004240 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.004249 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:08Z","lastTransitionTime":"2025-10-08T09:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.106475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.106802 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.106913 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.107017 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.107112 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:08Z","lastTransitionTime":"2025-10-08T09:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.210307 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.210434 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.210456 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.210482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.210500 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:08Z","lastTransitionTime":"2025-10-08T09:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.313626 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.313707 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.313731 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.313766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.313798 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:08Z","lastTransitionTime":"2025-10-08T09:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.416842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.416888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.416903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.416923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.416937 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:08Z","lastTransitionTime":"2025-10-08T09:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.452440 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:08 crc kubenswrapper[4744]: E1008 09:13:08.452666 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.520180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.520241 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.520263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.520292 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.520315 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:08Z","lastTransitionTime":"2025-10-08T09:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.623995 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.624062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.624084 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.624134 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.624157 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:08Z","lastTransitionTime":"2025-10-08T09:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.726344 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.726443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.726465 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.726491 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.726552 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:08Z","lastTransitionTime":"2025-10-08T09:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.829170 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.829234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.829256 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.829283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.829306 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:08Z","lastTransitionTime":"2025-10-08T09:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.932609 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.932656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.932669 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.932686 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:08 crc kubenswrapper[4744]: I1008 09:13:08.932698 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:08Z","lastTransitionTime":"2025-10-08T09:13:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.035860 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.035899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.035913 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.035931 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.035943 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:09Z","lastTransitionTime":"2025-10-08T09:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.139239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.139296 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.139310 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.139327 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.139338 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:09Z","lastTransitionTime":"2025-10-08T09:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.242911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.242966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.242983 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.243007 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.243029 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:09Z","lastTransitionTime":"2025-10-08T09:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.346033 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.346099 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.346115 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.346139 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.346160 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:09Z","lastTransitionTime":"2025-10-08T09:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.407594 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.407793 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.408426 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.408514 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.408773 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.408692855 +0000 UTC m=+148.656338134 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.408896 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.408937 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.408960 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.409059 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.409031025 +0000 UTC m=+148.656676294 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.409163 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.409263 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.409243661 +0000 UTC m=+148.656888930 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.409360 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.409558 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.409589 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.409615 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.409682 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.409662254 +0000 UTC m=+148.657307533 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.408946 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.410010 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.409972043 +0000 UTC m=+148.657617322 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.451043 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.451133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.451156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.451189 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.451211 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:09Z","lastTransitionTime":"2025-10-08T09:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.452219 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.452279 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.452324 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.452683 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.452824 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:09 crc kubenswrapper[4744]: E1008 09:13:09.452960 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.554522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.555021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.555209 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.555419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.555637 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:09Z","lastTransitionTime":"2025-10-08T09:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.658853 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.659039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.659142 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.659253 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.659344 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:09Z","lastTransitionTime":"2025-10-08T09:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.763137 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.763698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.763722 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.763749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.763768 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:09Z","lastTransitionTime":"2025-10-08T09:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.866836 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.866870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.866880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.866895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.866906 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:09Z","lastTransitionTime":"2025-10-08T09:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.970695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.970774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.970791 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.970814 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:09 crc kubenswrapper[4744]: I1008 09:13:09.970830 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:09Z","lastTransitionTime":"2025-10-08T09:13:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.073507 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.073543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.073553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.073569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.073580 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:10Z","lastTransitionTime":"2025-10-08T09:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.175880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.175926 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.175940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.175959 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.175974 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:10Z","lastTransitionTime":"2025-10-08T09:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.279649 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.279699 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.279717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.279740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.279761 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:10Z","lastTransitionTime":"2025-10-08T09:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.382974 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.383039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.383062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.383091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.383113 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:10Z","lastTransitionTime":"2025-10-08T09:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.452604 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:10 crc kubenswrapper[4744]: E1008 09:13:10.452765 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.485444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.485500 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.485513 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.485529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.485542 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:10Z","lastTransitionTime":"2025-10-08T09:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.588598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.588683 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.588694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.588706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.588716 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:10Z","lastTransitionTime":"2025-10-08T09:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.692487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.692579 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.692600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.692640 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.692681 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:10Z","lastTransitionTime":"2025-10-08T09:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.796334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.796431 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.796455 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.796483 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.796507 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:10Z","lastTransitionTime":"2025-10-08T09:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.899626 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.899666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.899676 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.899691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:10 crc kubenswrapper[4744]: I1008 09:13:10.899705 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:10Z","lastTransitionTime":"2025-10-08T09:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.002673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.002767 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.002787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.002817 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.002841 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:11Z","lastTransitionTime":"2025-10-08T09:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.106741 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.106816 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.106841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.106872 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.106893 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:11Z","lastTransitionTime":"2025-10-08T09:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.210749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.210808 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.210826 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.210850 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.210870 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:11Z","lastTransitionTime":"2025-10-08T09:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.319788 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.319854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.319873 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.319896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.319914 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:11Z","lastTransitionTime":"2025-10-08T09:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.423387 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.423438 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.423448 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.423463 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.423474 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:11Z","lastTransitionTime":"2025-10-08T09:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.452283 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:11 crc kubenswrapper[4744]: E1008 09:13:11.452489 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.452528 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:11 crc kubenswrapper[4744]: E1008 09:13:11.452753 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.453023 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:11 crc kubenswrapper[4744]: E1008 09:13:11.453496 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.526606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.526646 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.526655 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.526670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.526682 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:11Z","lastTransitionTime":"2025-10-08T09:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.629225 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.629266 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.629277 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.629295 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.629309 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:11Z","lastTransitionTime":"2025-10-08T09:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.731332 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.731399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.731412 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.731430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.731443 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:11Z","lastTransitionTime":"2025-10-08T09:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.833366 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.833421 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.833431 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.833444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.833452 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:11Z","lastTransitionTime":"2025-10-08T09:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.937319 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.937691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.937850 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.937995 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:11 crc kubenswrapper[4744]: I1008 09:13:11.938151 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:11Z","lastTransitionTime":"2025-10-08T09:13:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.042024 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.042146 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.042201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.042234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.042293 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:12Z","lastTransitionTime":"2025-10-08T09:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.145328 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.145386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.145399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.145415 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.145426 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:12Z","lastTransitionTime":"2025-10-08T09:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.250049 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.250412 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.250547 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.250656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.250737 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:12Z","lastTransitionTime":"2025-10-08T09:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.354715 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.354815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.354867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.354896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.354945 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:12Z","lastTransitionTime":"2025-10-08T09:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.423682 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.423820 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.423847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.423879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.423920 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T09:13:12Z","lastTransitionTime":"2025-10-08T09:13:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.455409 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:12 crc kubenswrapper[4744]: E1008 09:13:12.456366 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.459165 4744 scope.go:117] "RemoveContainer" containerID="00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357" Oct 08 09:13:12 crc kubenswrapper[4744]: E1008 09:13:12.459535 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.501564 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx"] Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.502008 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.511062 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.511129 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.511163 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.511836 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.582642 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jb62j" podStartSLOduration=67.582598985 podStartE2EDuration="1m7.582598985s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:12.567743436 +0000 UTC m=+87.815388725" watchObservedRunningTime="2025-10-08 09:13:12.582598985 +0000 UTC m=+87.830244264" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.620629 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=67.620609494 podStartE2EDuration="1m7.620609494s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:12.619651775 +0000 UTC m=+87.867297024" watchObservedRunningTime="2025-10-08 09:13:12.620609494 +0000 UTC m=+87.868254733" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.643316 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/22470030-6628-4e80-8d76-3a01d090ce30-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.643395 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22470030-6628-4e80-8d76-3a01d090ce30-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.643414 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/22470030-6628-4e80-8d76-3a01d090ce30-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.643434 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/22470030-6628-4e80-8d76-3a01d090ce30-service-ca\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.643465 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22470030-6628-4e80-8d76-3a01d090ce30-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.649266 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=33.649254589 podStartE2EDuration="33.649254589s" podCreationTimestamp="2025-10-08 09:12:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:12.634439892 +0000 UTC m=+87.882085141" watchObservedRunningTime="2025-10-08 09:13:12.649254589 +0000 UTC m=+87.896899828" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.699838 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-dldlq" podStartSLOduration=67.699817528 podStartE2EDuration="1m7.699817528s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:12.699264091 +0000 UTC m=+87.946909360" watchObservedRunningTime="2025-10-08 09:13:12.699817528 +0000 UTC m=+87.947462777" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.700344 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-j2skv" podStartSLOduration=67.700334294 podStartE2EDuration="1m7.700334294s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:12.680451812 +0000 UTC m=+87.928097081" watchObservedRunningTime="2025-10-08 09:13:12.700334294 +0000 UTC m=+87.947979553" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.734893 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=6.734875608 podStartE2EDuration="6.734875608s" podCreationTimestamp="2025-10-08 09:13:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:12.733879348 +0000 UTC m=+87.981524617" watchObservedRunningTime="2025-10-08 09:13:12.734875608 +0000 UTC m=+87.982520857" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.744648 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22470030-6628-4e80-8d76-3a01d090ce30-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.744705 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/22470030-6628-4e80-8d76-3a01d090ce30-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.744742 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/22470030-6628-4e80-8d76-3a01d090ce30-service-ca\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.744791 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22470030-6628-4e80-8d76-3a01d090ce30-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.744823 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/22470030-6628-4e80-8d76-3a01d090ce30-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.744920 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/22470030-6628-4e80-8d76-3a01d090ce30-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.745201 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/22470030-6628-4e80-8d76-3a01d090ce30-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.746212 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/22470030-6628-4e80-8d76-3a01d090ce30-service-ca\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.753070 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22470030-6628-4e80-8d76-3a01d090ce30-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.757628 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=61.757605015 podStartE2EDuration="1m1.757605015s" podCreationTimestamp="2025-10-08 09:12:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:12.757418799 +0000 UTC m=+88.005064058" watchObservedRunningTime="2025-10-08 09:13:12.757605015 +0000 UTC m=+88.005250274" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.768386 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22470030-6628-4e80-8d76-3a01d090ce30-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-r6bqx\" (UID: \"22470030-6628-4e80-8d76-3a01d090ce30\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.822443 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" Oct 08 09:13:12 crc kubenswrapper[4744]: W1008 09:13:12.841514 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22470030_6628_4e80_8d76_3a01d090ce30.slice/crio-501c2b72bc47a0fb19552364862dc5af37616c5b98d6ed8adfa5ae141e119e60 WatchSource:0}: Error finding container 501c2b72bc47a0fb19552364862dc5af37616c5b98d6ed8adfa5ae141e119e60: Status 404 returned error can't find the container with id 501c2b72bc47a0fb19552364862dc5af37616c5b98d6ed8adfa5ae141e119e60 Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.853556 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podStartSLOduration=67.853527314 podStartE2EDuration="1m7.853527314s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:12.852788252 +0000 UTC m=+88.100433491" watchObservedRunningTime="2025-10-08 09:13:12.853527314 +0000 UTC m=+88.101172553" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.853895 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-rgsf6" podStartSLOduration=67.853890605 podStartE2EDuration="1m7.853890605s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:12.831493918 +0000 UTC m=+88.079139167" watchObservedRunningTime="2025-10-08 09:13:12.853890605 +0000 UTC m=+88.101535844" Oct 08 09:13:12 crc kubenswrapper[4744]: I1008 09:13:12.894924 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-f6z7m" podStartSLOduration=67.894900645 podStartE2EDuration="1m7.894900645s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:12.89208398 +0000 UTC m=+88.139729269" watchObservedRunningTime="2025-10-08 09:13:12.894900645 +0000 UTC m=+88.142545894" Oct 08 09:13:13 crc kubenswrapper[4744]: I1008 09:13:13.055028 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" event={"ID":"22470030-6628-4e80-8d76-3a01d090ce30","Type":"ContainerStarted","Data":"16da8039ba5f0c65f93549afd2c015cc37d1063ff0e7162c70a7a06ac7cce618"} Oct 08 09:13:13 crc kubenswrapper[4744]: I1008 09:13:13.055082 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" event={"ID":"22470030-6628-4e80-8d76-3a01d090ce30","Type":"ContainerStarted","Data":"501c2b72bc47a0fb19552364862dc5af37616c5b98d6ed8adfa5ae141e119e60"} Oct 08 09:13:13 crc kubenswrapper[4744]: I1008 09:13:13.083198 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-r6bqx" podStartSLOduration=68.083168946 podStartE2EDuration="1m8.083168946s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:13.080804894 +0000 UTC m=+88.328450143" watchObservedRunningTime="2025-10-08 09:13:13.083168946 +0000 UTC m=+88.330814225" Oct 08 09:13:13 crc kubenswrapper[4744]: I1008 09:13:13.453219 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:13 crc kubenswrapper[4744]: I1008 09:13:13.453351 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:13 crc kubenswrapper[4744]: E1008 09:13:13.453424 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:13 crc kubenswrapper[4744]: E1008 09:13:13.453589 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:13 crc kubenswrapper[4744]: I1008 09:13:13.453796 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:13 crc kubenswrapper[4744]: E1008 09:13:13.453878 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:14 crc kubenswrapper[4744]: I1008 09:13:14.452421 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:14 crc kubenswrapper[4744]: E1008 09:13:14.452772 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:14 crc kubenswrapper[4744]: I1008 09:13:14.467364 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 08 09:13:15 crc kubenswrapper[4744]: I1008 09:13:15.452235 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:15 crc kubenswrapper[4744]: I1008 09:13:15.452182 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:15 crc kubenswrapper[4744]: I1008 09:13:15.452700 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:15 crc kubenswrapper[4744]: E1008 09:13:15.454264 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:15 crc kubenswrapper[4744]: E1008 09:13:15.454540 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:15 crc kubenswrapper[4744]: E1008 09:13:15.454635 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:15 crc kubenswrapper[4744]: I1008 09:13:15.472320 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=1.472297084 podStartE2EDuration="1.472297084s" podCreationTimestamp="2025-10-08 09:13:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:15.47150851 +0000 UTC m=+90.719153779" watchObservedRunningTime="2025-10-08 09:13:15.472297084 +0000 UTC m=+90.719942353" Oct 08 09:13:16 crc kubenswrapper[4744]: I1008 09:13:16.452083 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:16 crc kubenswrapper[4744]: E1008 09:13:16.452216 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:17 crc kubenswrapper[4744]: I1008 09:13:17.452863 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:17 crc kubenswrapper[4744]: I1008 09:13:17.452964 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:17 crc kubenswrapper[4744]: I1008 09:13:17.453092 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:17 crc kubenswrapper[4744]: E1008 09:13:17.453818 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:17 crc kubenswrapper[4744]: E1008 09:13:17.453489 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:17 crc kubenswrapper[4744]: E1008 09:13:17.453264 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:18 crc kubenswrapper[4744]: I1008 09:13:18.452756 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:18 crc kubenswrapper[4744]: E1008 09:13:18.453167 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:19 crc kubenswrapper[4744]: I1008 09:13:19.452992 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:19 crc kubenswrapper[4744]: I1008 09:13:19.453003 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:19 crc kubenswrapper[4744]: E1008 09:13:19.453147 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:19 crc kubenswrapper[4744]: I1008 09:13:19.453021 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:19 crc kubenswrapper[4744]: E1008 09:13:19.453226 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:19 crc kubenswrapper[4744]: E1008 09:13:19.453501 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:20 crc kubenswrapper[4744]: I1008 09:13:20.452741 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:20 crc kubenswrapper[4744]: E1008 09:13:20.452940 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:21 crc kubenswrapper[4744]: I1008 09:13:21.452346 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:21 crc kubenswrapper[4744]: I1008 09:13:21.452364 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:21 crc kubenswrapper[4744]: E1008 09:13:21.453315 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:21 crc kubenswrapper[4744]: E1008 09:13:21.453152 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:21 crc kubenswrapper[4744]: I1008 09:13:21.452538 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:21 crc kubenswrapper[4744]: E1008 09:13:21.453451 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:22 crc kubenswrapper[4744]: I1008 09:13:22.452592 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:22 crc kubenswrapper[4744]: E1008 09:13:22.452819 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:23 crc kubenswrapper[4744]: I1008 09:13:23.452410 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:23 crc kubenswrapper[4744]: I1008 09:13:23.452490 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:23 crc kubenswrapper[4744]: E1008 09:13:23.452604 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:23 crc kubenswrapper[4744]: I1008 09:13:23.452642 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:23 crc kubenswrapper[4744]: E1008 09:13:23.452802 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:23 crc kubenswrapper[4744]: E1008 09:13:23.452909 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:23 crc kubenswrapper[4744]: I1008 09:13:23.764833 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:23 crc kubenswrapper[4744]: E1008 09:13:23.765095 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:13:23 crc kubenswrapper[4744]: E1008 09:13:23.765179 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs podName:be364129-0f0b-41d4-b5f8-1b1def9a1c38 nodeName:}" failed. No retries permitted until 2025-10-08 09:14:27.765155667 +0000 UTC m=+163.012800946 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs") pod "network-metrics-daemon-cdcbz" (UID: "be364129-0f0b-41d4-b5f8-1b1def9a1c38") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 09:13:24 crc kubenswrapper[4744]: I1008 09:13:24.452603 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:24 crc kubenswrapper[4744]: E1008 09:13:24.452915 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:25 crc kubenswrapper[4744]: I1008 09:13:25.452307 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:25 crc kubenswrapper[4744]: I1008 09:13:25.452314 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:25 crc kubenswrapper[4744]: I1008 09:13:25.452484 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:25 crc kubenswrapper[4744]: E1008 09:13:25.454662 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:25 crc kubenswrapper[4744]: E1008 09:13:25.454799 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:25 crc kubenswrapper[4744]: E1008 09:13:25.454900 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:26 crc kubenswrapper[4744]: I1008 09:13:26.452973 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:26 crc kubenswrapper[4744]: E1008 09:13:26.453148 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:27 crc kubenswrapper[4744]: I1008 09:13:27.452501 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:27 crc kubenswrapper[4744]: I1008 09:13:27.452542 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:27 crc kubenswrapper[4744]: E1008 09:13:27.452706 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:27 crc kubenswrapper[4744]: I1008 09:13:27.453429 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:27 crc kubenswrapper[4744]: E1008 09:13:27.453484 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:27 crc kubenswrapper[4744]: I1008 09:13:27.453525 4744 scope.go:117] "RemoveContainer" containerID="00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357" Oct 08 09:13:27 crc kubenswrapper[4744]: E1008 09:13:27.453820 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:27 crc kubenswrapper[4744]: E1008 09:13:27.453858 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tmmnj_openshift-ovn-kubernetes(5a9bf215-f18b-4732-b12c-0355a3a0a15f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" Oct 08 09:13:28 crc kubenswrapper[4744]: I1008 09:13:28.452667 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:28 crc kubenswrapper[4744]: E1008 09:13:28.452821 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:29 crc kubenswrapper[4744]: I1008 09:13:29.452880 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:29 crc kubenswrapper[4744]: I1008 09:13:29.452994 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:29 crc kubenswrapper[4744]: I1008 09:13:29.453097 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:29 crc kubenswrapper[4744]: E1008 09:13:29.453108 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:29 crc kubenswrapper[4744]: E1008 09:13:29.453344 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:29 crc kubenswrapper[4744]: E1008 09:13:29.453479 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:30 crc kubenswrapper[4744]: I1008 09:13:30.452061 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:30 crc kubenswrapper[4744]: E1008 09:13:30.452222 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:31 crc kubenswrapper[4744]: I1008 09:13:31.452456 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:31 crc kubenswrapper[4744]: I1008 09:13:31.452505 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:31 crc kubenswrapper[4744]: E1008 09:13:31.452619 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:31 crc kubenswrapper[4744]: I1008 09:13:31.452887 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:31 crc kubenswrapper[4744]: E1008 09:13:31.453025 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:31 crc kubenswrapper[4744]: E1008 09:13:31.453464 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:32 crc kubenswrapper[4744]: I1008 09:13:32.452328 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:32 crc kubenswrapper[4744]: E1008 09:13:32.452606 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:33 crc kubenswrapper[4744]: I1008 09:13:33.452773 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:33 crc kubenswrapper[4744]: I1008 09:13:33.452915 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:33 crc kubenswrapper[4744]: I1008 09:13:33.452951 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:33 crc kubenswrapper[4744]: E1008 09:13:33.454586 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:33 crc kubenswrapper[4744]: E1008 09:13:33.454695 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:33 crc kubenswrapper[4744]: E1008 09:13:33.454822 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:34 crc kubenswrapper[4744]: I1008 09:13:34.452046 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:34 crc kubenswrapper[4744]: E1008 09:13:34.452214 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:35 crc kubenswrapper[4744]: I1008 09:13:35.452234 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:35 crc kubenswrapper[4744]: I1008 09:13:35.452227 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:35 crc kubenswrapper[4744]: I1008 09:13:35.452316 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:35 crc kubenswrapper[4744]: E1008 09:13:35.454930 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:35 crc kubenswrapper[4744]: E1008 09:13:35.455130 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:35 crc kubenswrapper[4744]: E1008 09:13:35.455364 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:36 crc kubenswrapper[4744]: I1008 09:13:36.452513 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:36 crc kubenswrapper[4744]: E1008 09:13:36.452715 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:37 crc kubenswrapper[4744]: I1008 09:13:37.452412 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:37 crc kubenswrapper[4744]: I1008 09:13:37.452416 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:37 crc kubenswrapper[4744]: I1008 09:13:37.453778 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:37 crc kubenswrapper[4744]: E1008 09:13:37.453893 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:37 crc kubenswrapper[4744]: E1008 09:13:37.454304 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:37 crc kubenswrapper[4744]: E1008 09:13:37.454401 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:38 crc kubenswrapper[4744]: I1008 09:13:38.452853 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:38 crc kubenswrapper[4744]: E1008 09:13:38.453031 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:39 crc kubenswrapper[4744]: I1008 09:13:39.453020 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:39 crc kubenswrapper[4744]: E1008 09:13:39.453265 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:39 crc kubenswrapper[4744]: I1008 09:13:39.453647 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:39 crc kubenswrapper[4744]: E1008 09:13:39.453777 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:39 crc kubenswrapper[4744]: I1008 09:13:39.454027 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:39 crc kubenswrapper[4744]: E1008 09:13:39.454154 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:40 crc kubenswrapper[4744]: I1008 09:13:40.145805 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dldlq_7ea48507-1f33-46fc-ab34-c6e083dd86a9/kube-multus/1.log" Oct 08 09:13:40 crc kubenswrapper[4744]: I1008 09:13:40.146288 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dldlq_7ea48507-1f33-46fc-ab34-c6e083dd86a9/kube-multus/0.log" Oct 08 09:13:40 crc kubenswrapper[4744]: I1008 09:13:40.146328 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ea48507-1f33-46fc-ab34-c6e083dd86a9" containerID="6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75" exitCode=1 Oct 08 09:13:40 crc kubenswrapper[4744]: I1008 09:13:40.146415 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dldlq" event={"ID":"7ea48507-1f33-46fc-ab34-c6e083dd86a9","Type":"ContainerDied","Data":"6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75"} Oct 08 09:13:40 crc kubenswrapper[4744]: I1008 09:13:40.146454 4744 scope.go:117] "RemoveContainer" containerID="706b44a7205f6b25d4e02d69139f294968e14c4b0dd8c21587e195f6a56ecd6f" Oct 08 09:13:40 crc kubenswrapper[4744]: I1008 09:13:40.146816 4744 scope.go:117] "RemoveContainer" containerID="6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75" Oct 08 09:13:40 crc kubenswrapper[4744]: E1008 09:13:40.146972 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-dldlq_openshift-multus(7ea48507-1f33-46fc-ab34-c6e083dd86a9)\"" pod="openshift-multus/multus-dldlq" podUID="7ea48507-1f33-46fc-ab34-c6e083dd86a9" Oct 08 09:13:40 crc kubenswrapper[4744]: I1008 09:13:40.452034 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:40 crc kubenswrapper[4744]: E1008 09:13:40.452160 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:41 crc kubenswrapper[4744]: I1008 09:13:41.152109 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dldlq_7ea48507-1f33-46fc-ab34-c6e083dd86a9/kube-multus/1.log" Oct 08 09:13:41 crc kubenswrapper[4744]: I1008 09:13:41.453232 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:41 crc kubenswrapper[4744]: I1008 09:13:41.453334 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:41 crc kubenswrapper[4744]: E1008 09:13:41.454004 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:41 crc kubenswrapper[4744]: I1008 09:13:41.454029 4744 scope.go:117] "RemoveContainer" containerID="00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357" Oct 08 09:13:41 crc kubenswrapper[4744]: E1008 09:13:41.454213 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:41 crc kubenswrapper[4744]: I1008 09:13:41.454444 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:41 crc kubenswrapper[4744]: E1008 09:13:41.454564 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:42 crc kubenswrapper[4744]: I1008 09:13:42.157439 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/3.log" Oct 08 09:13:42 crc kubenswrapper[4744]: I1008 09:13:42.159583 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerStarted","Data":"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a"} Oct 08 09:13:42 crc kubenswrapper[4744]: I1008 09:13:42.160070 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:13:42 crc kubenswrapper[4744]: I1008 09:13:42.301027 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podStartSLOduration=97.301002293 podStartE2EDuration="1m37.301002293s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:13:42.190217973 +0000 UTC m=+117.437863222" watchObservedRunningTime="2025-10-08 09:13:42.301002293 +0000 UTC m=+117.548647552" Oct 08 09:13:42 crc kubenswrapper[4744]: I1008 09:13:42.302415 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-cdcbz"] Oct 08 09:13:42 crc kubenswrapper[4744]: I1008 09:13:42.302548 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:42 crc kubenswrapper[4744]: E1008 09:13:42.302670 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:42 crc kubenswrapper[4744]: I1008 09:13:42.452853 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:42 crc kubenswrapper[4744]: E1008 09:13:42.453050 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:43 crc kubenswrapper[4744]: I1008 09:13:43.452514 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:43 crc kubenswrapper[4744]: I1008 09:13:43.452547 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:43 crc kubenswrapper[4744]: E1008 09:13:43.452697 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:43 crc kubenswrapper[4744]: E1008 09:13:43.452788 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:44 crc kubenswrapper[4744]: I1008 09:13:44.452480 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:44 crc kubenswrapper[4744]: I1008 09:13:44.452569 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:44 crc kubenswrapper[4744]: E1008 09:13:44.452712 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:44 crc kubenswrapper[4744]: E1008 09:13:44.452848 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:45 crc kubenswrapper[4744]: E1008 09:13:45.440413 4744 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 08 09:13:45 crc kubenswrapper[4744]: I1008 09:13:45.452671 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:45 crc kubenswrapper[4744]: I1008 09:13:45.452731 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:45 crc kubenswrapper[4744]: E1008 09:13:45.452811 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:45 crc kubenswrapper[4744]: E1008 09:13:45.452927 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:45 crc kubenswrapper[4744]: E1008 09:13:45.553090 4744 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 09:13:46 crc kubenswrapper[4744]: I1008 09:13:46.452359 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:46 crc kubenswrapper[4744]: I1008 09:13:46.452359 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:46 crc kubenswrapper[4744]: E1008 09:13:46.452494 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:46 crc kubenswrapper[4744]: E1008 09:13:46.452554 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:47 crc kubenswrapper[4744]: I1008 09:13:47.452620 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:47 crc kubenswrapper[4744]: E1008 09:13:47.452940 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:47 crc kubenswrapper[4744]: I1008 09:13:47.453010 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:47 crc kubenswrapper[4744]: E1008 09:13:47.453206 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:48 crc kubenswrapper[4744]: I1008 09:13:48.452812 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:48 crc kubenswrapper[4744]: I1008 09:13:48.452812 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:48 crc kubenswrapper[4744]: E1008 09:13:48.453040 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:48 crc kubenswrapper[4744]: E1008 09:13:48.452936 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:49 crc kubenswrapper[4744]: I1008 09:13:49.452839 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:49 crc kubenswrapper[4744]: E1008 09:13:49.452984 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:49 crc kubenswrapper[4744]: I1008 09:13:49.453061 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:49 crc kubenswrapper[4744]: E1008 09:13:49.453207 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:50 crc kubenswrapper[4744]: I1008 09:13:50.452312 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:50 crc kubenswrapper[4744]: I1008 09:13:50.452412 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:50 crc kubenswrapper[4744]: E1008 09:13:50.452506 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:50 crc kubenswrapper[4744]: E1008 09:13:50.452582 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:50 crc kubenswrapper[4744]: E1008 09:13:50.554832 4744 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 09:13:51 crc kubenswrapper[4744]: I1008 09:13:51.452860 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:51 crc kubenswrapper[4744]: I1008 09:13:51.452978 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:51 crc kubenswrapper[4744]: E1008 09:13:51.453072 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:51 crc kubenswrapper[4744]: E1008 09:13:51.453132 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:52 crc kubenswrapper[4744]: I1008 09:13:52.452911 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:52 crc kubenswrapper[4744]: I1008 09:13:52.452940 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:52 crc kubenswrapper[4744]: E1008 09:13:52.453036 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:52 crc kubenswrapper[4744]: E1008 09:13:52.453411 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:53 crc kubenswrapper[4744]: I1008 09:13:53.452944 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:53 crc kubenswrapper[4744]: E1008 09:13:53.453123 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:53 crc kubenswrapper[4744]: I1008 09:13:53.454093 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:53 crc kubenswrapper[4744]: E1008 09:13:53.454401 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:54 crc kubenswrapper[4744]: I1008 09:13:54.452577 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:54 crc kubenswrapper[4744]: I1008 09:13:54.452658 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:54 crc kubenswrapper[4744]: I1008 09:13:54.453117 4744 scope.go:117] "RemoveContainer" containerID="6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75" Oct 08 09:13:54 crc kubenswrapper[4744]: E1008 09:13:54.453147 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:54 crc kubenswrapper[4744]: E1008 09:13:54.453182 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:55 crc kubenswrapper[4744]: I1008 09:13:55.208564 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dldlq_7ea48507-1f33-46fc-ab34-c6e083dd86a9/kube-multus/1.log" Oct 08 09:13:55 crc kubenswrapper[4744]: I1008 09:13:55.208986 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dldlq" event={"ID":"7ea48507-1f33-46fc-ab34-c6e083dd86a9","Type":"ContainerStarted","Data":"94b68c9a4976717260751eead14fea87839d1758f32a92629bc6776f522afb3e"} Oct 08 09:13:55 crc kubenswrapper[4744]: I1008 09:13:55.453069 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:55 crc kubenswrapper[4744]: E1008 09:13:55.453235 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:55 crc kubenswrapper[4744]: I1008 09:13:55.454134 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:55 crc kubenswrapper[4744]: E1008 09:13:55.454296 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:55 crc kubenswrapper[4744]: E1008 09:13:55.556020 4744 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 09:13:56 crc kubenswrapper[4744]: I1008 09:13:56.453077 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:56 crc kubenswrapper[4744]: I1008 09:13:56.453077 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:56 crc kubenswrapper[4744]: E1008 09:13:56.453287 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:56 crc kubenswrapper[4744]: E1008 09:13:56.453562 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:57 crc kubenswrapper[4744]: I1008 09:13:57.453009 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:57 crc kubenswrapper[4744]: E1008 09:13:57.453135 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:13:57 crc kubenswrapper[4744]: I1008 09:13:57.453009 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:57 crc kubenswrapper[4744]: E1008 09:13:57.453243 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:58 crc kubenswrapper[4744]: I1008 09:13:58.453130 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:13:58 crc kubenswrapper[4744]: I1008 09:13:58.453142 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:13:58 crc kubenswrapper[4744]: E1008 09:13:58.455131 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:13:58 crc kubenswrapper[4744]: E1008 09:13:58.455272 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:13:59 crc kubenswrapper[4744]: I1008 09:13:59.452730 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:13:59 crc kubenswrapper[4744]: E1008 09:13:59.452880 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 09:13:59 crc kubenswrapper[4744]: I1008 09:13:59.452893 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:13:59 crc kubenswrapper[4744]: E1008 09:13:59.452982 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 09:14:00 crc kubenswrapper[4744]: I1008 09:14:00.452132 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:14:00 crc kubenswrapper[4744]: I1008 09:14:00.452170 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:14:00 crc kubenswrapper[4744]: E1008 09:14:00.452249 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cdcbz" podUID="be364129-0f0b-41d4-b5f8-1b1def9a1c38" Oct 08 09:14:00 crc kubenswrapper[4744]: E1008 09:14:00.452503 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 09:14:01 crc kubenswrapper[4744]: I1008 09:14:01.358901 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:14:01 crc kubenswrapper[4744]: I1008 09:14:01.452508 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:14:01 crc kubenswrapper[4744]: I1008 09:14:01.452617 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:14:01 crc kubenswrapper[4744]: I1008 09:14:01.455209 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 08 09:14:01 crc kubenswrapper[4744]: I1008 09:14:01.455523 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 08 09:14:02 crc kubenswrapper[4744]: I1008 09:14:02.453040 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:14:02 crc kubenswrapper[4744]: I1008 09:14:02.453075 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:14:02 crc kubenswrapper[4744]: I1008 09:14:02.459025 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 08 09:14:02 crc kubenswrapper[4744]: I1008 09:14:02.459608 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 08 09:14:02 crc kubenswrapper[4744]: I1008 09:14:02.460215 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 08 09:14:02 crc kubenswrapper[4744]: I1008 09:14:02.460685 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.278431 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.360764 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-b5q8z"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.361831 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.362005 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.363033 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.363224 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.364426 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.367203 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.370198 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.370555 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.370718 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.370838 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.370945 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.378237 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.380116 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fjhhp"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.380787 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.381324 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.381512 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.382044 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.383675 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.383677 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.385123 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.385695 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.386166 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.386839 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.387402 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-5j25g"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.387744 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.388325 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fl2qz"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.388846 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.391899 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.392303 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.392499 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-n26p9"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.398923 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.399152 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.405762 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.405829 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.406001 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.406454 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.406875 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.407330 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.407482 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.407762 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.407834 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.408036 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.408148 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.408217 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.408419 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.408528 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.408848 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.409029 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.409610 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.411411 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.411522 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.413810 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.416949 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.417432 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.418025 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.418129 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-n26p9" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.418597 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.419099 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.426759 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.427283 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.427464 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.427867 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.429655 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.430144 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.431965 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.432092 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.432422 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.432834 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.434196 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.434397 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pjz2t"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.434765 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.435003 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.435286 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.435310 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.439024 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vkldx"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.439423 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.439934 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.444619 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8h78h"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.445452 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.447847 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zc56g"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.448007 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.448385 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.449087 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.452189 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5lxjl"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.452588 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.452909 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x97sx"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.453194 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.453677 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.453932 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.473959 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.474517 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-8whxk"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.474786 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.475165 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cf7sw"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.475454 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zklvg"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.475807 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.476273 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.476522 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dr8c5"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.476843 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-f8825"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.477180 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-jgg9g"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.477739 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.477857 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jgg9g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.478003 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.478228 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.478336 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.478694 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.499125 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.500312 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.500492 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8whxk" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.500739 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.501150 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.505686 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.506008 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.508340 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.508382 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.509917 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.510520 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.511389 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.512056 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.512495 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-dr8c5" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.512822 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.512852 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.512970 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.513055 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.513192 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.513011 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.477760 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hflmk"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.514995 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.529296 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.530047 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.530546 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/800b4855-f900-4968-9a86-fad43cec8fa6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-99rp7\" (UID: \"800b4855-f900-4968-9a86-fad43cec8fa6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.530643 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtjgm\" (UniqueName: \"kubernetes.io/projected/10339d22-99de-4f28-98a2-f3549c0f4c13-kube-api-access-xtjgm\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.530720 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1459a017-1dc3-4189-b81f-0f4dfcee04f4-profile-collector-cert\") pod \"catalog-operator-68c6474976-tfgxd\" (UID: \"1459a017-1dc3-4189-b81f-0f4dfcee04f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.530791 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbc99729-cada-43b3-be3e-528a87db765e-serving-cert\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.530863 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc63fe57-540e-4c29-8fc6-155d220f2e9e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m4wbt\" (UID: \"fc63fe57-540e-4c29-8fc6-155d220f2e9e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.531006 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/10339d22-99de-4f28-98a2-f3549c0f4c13-etcd-client\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.531073 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/23ba3d67-97df-4a13-8603-4c0efa967852-metrics-certs\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.531141 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05957de0-ed69-46e2-aab9-d146cfcbbb90-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-54nth\" (UID: \"05957de0-ed69-46e2-aab9-d146cfcbbb90\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.531207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/800b4855-f900-4968-9a86-fad43cec8fa6-config\") pod \"kube-controller-manager-operator-78b949d7b-99rp7\" (UID: \"800b4855-f900-4968-9a86-fad43cec8fa6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.531276 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-etcd-serving-ca\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.531347 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q76n\" (UniqueName: \"kubernetes.io/projected/81cf0144-e970-4fe2-8663-15b006dc1b91-kube-api-access-2q76n\") pod \"control-plane-machine-set-operator-78cbb6b69f-5nnfc\" (UID: \"81cf0144-e970-4fe2-8663-15b006dc1b91\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.531451 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc63fe57-540e-4c29-8fc6-155d220f2e9e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m4wbt\" (UID: \"fc63fe57-540e-4c29-8fc6-155d220f2e9e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.531784 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.531522 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/62b2562b-f25e-4276-bab1-0b366d60c231-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n9d98\" (UID: \"62b2562b-f25e-4276-bab1-0b366d60c231\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532623 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05957de0-ed69-46e2-aab9-d146cfcbbb90-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-54nth\" (UID: \"05957de0-ed69-46e2-aab9-d146cfcbbb90\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532648 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc63fe57-540e-4c29-8fc6-155d220f2e9e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m4wbt\" (UID: \"fc63fe57-540e-4c29-8fc6-155d220f2e9e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532670 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bacbc855-fe99-4250-9da3-b64e95ac7c67-secret-volume\") pod \"collect-profiles-29331900-b46lv\" (UID: \"bacbc855-fe99-4250-9da3-b64e95ac7c67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532687 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rzws\" (UniqueName: \"kubernetes.io/projected/dc5708d4-d312-48a6-931a-b09f0c1dd569-kube-api-access-9rzws\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532705 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhrvh\" (UniqueName: \"kubernetes.io/projected/18e8bf6d-bea3-435c-9e4b-4d7a287f0c83-kube-api-access-zhrvh\") pod \"machine-approver-56656f9798-nn48b\" (UID: \"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532720 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc5708d4-d312-48a6-931a-b09f0c1dd569-service-ca-bundle\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532734 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff9cb19e-ad39-4636-b302-57b5f39b0072-client-ca\") pod \"route-controller-manager-6576b87f9c-hns6n\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532752 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7221db0-3f54-4e07-95ec-56c4fb95aecc-config\") pod \"machine-api-operator-5694c8668f-b5q8z\" (UID: \"a7221db0-3f54-4e07-95ec-56c4fb95aecc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532782 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/800b4855-f900-4968-9a86-fad43cec8fa6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-99rp7\" (UID: \"800b4855-f900-4968-9a86-fad43cec8fa6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532798 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdnxn\" (UniqueName: \"kubernetes.io/projected/bacbc855-fe99-4250-9da3-b64e95ac7c67-kube-api-access-wdnxn\") pod \"collect-profiles-29331900-b46lv\" (UID: \"bacbc855-fe99-4250-9da3-b64e95ac7c67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532816 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-image-import-ca\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/23ba3d67-97df-4a13-8603-4c0efa967852-default-certificate\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532851 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/81cf0144-e970-4fe2-8663-15b006dc1b91-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5nnfc\" (UID: \"81cf0144-e970-4fe2-8663-15b006dc1b91\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532871 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a7221db0-3f54-4e07-95ec-56c4fb95aecc-images\") pod \"machine-api-operator-5694c8668f-b5q8z\" (UID: \"a7221db0-3f54-4e07-95ec-56c4fb95aecc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532887 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p59xk\" (UniqueName: \"kubernetes.io/projected/05957de0-ed69-46e2-aab9-d146cfcbbb90-kube-api-access-p59xk\") pod \"kube-storage-version-migrator-operator-b67b599dd-54nth\" (UID: \"05957de0-ed69-46e2-aab9-d146cfcbbb90\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532906 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a7221db0-3f54-4e07-95ec-56c4fb95aecc-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-b5q8z\" (UID: \"a7221db0-3f54-4e07-95ec-56c4fb95aecc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532920 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fbc99729-cada-43b3-be3e-528a87db765e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532935 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1459a017-1dc3-4189-b81f-0f4dfcee04f4-srv-cert\") pod \"catalog-operator-68c6474976-tfgxd\" (UID: \"1459a017-1dc3-4189-b81f-0f4dfcee04f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532955 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-audit\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532977 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdql7\" (UniqueName: \"kubernetes.io/projected/8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a-kube-api-access-vdql7\") pod \"multus-admission-controller-857f4d67dd-n26p9\" (UID: \"8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n26p9" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532992 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/10339d22-99de-4f28-98a2-f3549c0f4c13-encryption-config\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533007 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lcpx\" (UniqueName: \"kubernetes.io/projected/23ba3d67-97df-4a13-8603-4c0efa967852-kube-api-access-9lcpx\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533024 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18e8bf6d-bea3-435c-9e4b-4d7a287f0c83-config\") pod \"machine-approver-56656f9798-nn48b\" (UID: \"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533038 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-n26p9\" (UID: \"8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n26p9" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533054 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fbc99729-cada-43b3-be3e-528a87db765e-audit-dir\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533092 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/18e8bf6d-bea3-435c-9e4b-4d7a287f0c83-machine-approver-tls\") pod \"machine-approver-56656f9798-nn48b\" (UID: \"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533107 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc5708d4-d312-48a6-931a-b09f0c1dd569-serving-cert\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533125 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svjqj\" (UniqueName: \"kubernetes.io/projected/ff9cb19e-ad39-4636-b302-57b5f39b0072-kube-api-access-svjqj\") pod \"route-controller-manager-6576b87f9c-hns6n\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533142 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/10339d22-99de-4f28-98a2-f3549c0f4c13-node-pullsecrets\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533158 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62b2562b-f25e-4276-bab1-0b366d60c231-trusted-ca\") pod \"ingress-operator-5b745b69d9-n9d98\" (UID: \"62b2562b-f25e-4276-bab1-0b366d60c231\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533180 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff9cb19e-ad39-4636-b302-57b5f39b0072-serving-cert\") pod \"route-controller-manager-6576b87f9c-hns6n\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533195 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r89m8\" (UniqueName: \"kubernetes.io/projected/1459a017-1dc3-4189-b81f-0f4dfcee04f4-kube-api-access-r89m8\") pod \"catalog-operator-68c6474976-tfgxd\" (UID: \"1459a017-1dc3-4189-b81f-0f4dfcee04f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533213 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10339d22-99de-4f28-98a2-f3549c0f4c13-serving-cert\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533229 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bacbc855-fe99-4250-9da3-b64e95ac7c67-config-volume\") pod \"collect-profiles-29331900-b46lv\" (UID: \"bacbc855-fe99-4250-9da3-b64e95ac7c67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533246 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fbc99729-cada-43b3-be3e-528a87db765e-encryption-config\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533264 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-config\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533278 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/23ba3d67-97df-4a13-8603-4c0efa967852-stats-auth\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533291 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/62b2562b-f25e-4276-bab1-0b366d60c231-metrics-tls\") pod \"ingress-operator-5b745b69d9-n9d98\" (UID: \"62b2562b-f25e-4276-bab1-0b366d60c231\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533305 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4vh5\" (UniqueName: \"kubernetes.io/projected/62b2562b-f25e-4276-bab1-0b366d60c231-kube-api-access-f4vh5\") pod \"ingress-operator-5b745b69d9-n9d98\" (UID: \"62b2562b-f25e-4276-bab1-0b366d60c231\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533321 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fbc99729-cada-43b3-be3e-528a87db765e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533343 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18e8bf6d-bea3-435c-9e4b-4d7a287f0c83-auth-proxy-config\") pod \"machine-approver-56656f9798-nn48b\" (UID: \"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533358 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fbc99729-cada-43b3-be3e-528a87db765e-etcd-client\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533390 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26t5q\" (UniqueName: \"kubernetes.io/projected/fbc99729-cada-43b3-be3e-528a87db765e-kube-api-access-26t5q\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533403 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff9cb19e-ad39-4636-b302-57b5f39b0072-config\") pod \"route-controller-manager-6576b87f9c-hns6n\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533421 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fbc99729-cada-43b3-be3e-528a87db765e-audit-policies\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533435 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc5708d4-d312-48a6-931a-b09f0c1dd569-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533449 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533480 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc5708d4-d312-48a6-931a-b09f0c1dd569-config\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533496 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvxs2\" (UniqueName: \"kubernetes.io/projected/a7221db0-3f54-4e07-95ec-56c4fb95aecc-kube-api-access-dvxs2\") pod \"machine-api-operator-5694c8668f-b5q8z\" (UID: \"a7221db0-3f54-4e07-95ec-56c4fb95aecc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533511 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/10339d22-99de-4f28-98a2-f3549c0f4c13-audit-dir\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.533546 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/23ba3d67-97df-4a13-8603-4c0efa967852-service-ca-bundle\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.531935 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.531965 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.531978 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532000 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532019 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532060 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.539531 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.532148 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.534995 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.535037 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.535074 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.535105 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.535136 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.535361 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.535434 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.535484 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.536187 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.536339 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.536425 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.536504 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.536558 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.536709 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.536740 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.536778 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.536816 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.536845 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.536872 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.538422 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.539968 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.541191 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.541322 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.540865 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.542634 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.543236 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.535639 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.537029 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.535563 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-5kssv"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.544955 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.546233 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.546433 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.546277 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.545122 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.545281 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.547430 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.547743 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.548049 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-b5q8z"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.548064 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9knh2"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.548807 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.549002 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.549134 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.549264 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.550010 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.551627 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.552451 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.553061 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.554575 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.555611 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.555692 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.558683 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fl2qz"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.561421 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8whxk"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.561462 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.569876 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.572355 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.572435 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8h78h"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.572948 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.575974 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.576028 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fjhhp"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.580150 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.582699 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.584864 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-tbc7j"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.585481 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-tbc7j" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.586865 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.587675 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-fcggf"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.594859 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.594967 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fcggf" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.607769 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-n26p9"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.609784 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.612577 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pjz2t"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.614148 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.617654 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cf7sw"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.620448 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.622256 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.625228 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.630151 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.632077 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zklvg"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.640909 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/23ba3d67-97df-4a13-8603-4c0efa967852-default-certificate\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.640950 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.640966 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/81cf0144-e970-4fe2-8663-15b006dc1b91-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5nnfc\" (UID: \"81cf0144-e970-4fe2-8663-15b006dc1b91\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.640987 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vkldx"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.640989 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7221db0-3f54-4e07-95ec-56c4fb95aecc-config\") pod \"machine-api-operator-5694c8668f-b5q8z\" (UID: \"a7221db0-3f54-4e07-95ec-56c4fb95aecc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.640998 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5lxjl"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641008 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/800b4855-f900-4968-9a86-fad43cec8fa6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-99rp7\" (UID: \"800b4855-f900-4968-9a86-fad43cec8fa6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641025 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdnxn\" (UniqueName: \"kubernetes.io/projected/bacbc855-fe99-4250-9da3-b64e95ac7c67-kube-api-access-wdnxn\") pod \"collect-profiles-29331900-b46lv\" (UID: \"bacbc855-fe99-4250-9da3-b64e95ac7c67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641040 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-image-import-ca\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641078 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a7221db0-3f54-4e07-95ec-56c4fb95aecc-images\") pod \"machine-api-operator-5694c8668f-b5q8z\" (UID: \"a7221db0-3f54-4e07-95ec-56c4fb95aecc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641096 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p59xk\" (UniqueName: \"kubernetes.io/projected/05957de0-ed69-46e2-aab9-d146cfcbbb90-kube-api-access-p59xk\") pod \"kube-storage-version-migrator-operator-b67b599dd-54nth\" (UID: \"05957de0-ed69-46e2-aab9-d146cfcbbb90\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641119 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xvzs\" (UniqueName: \"kubernetes.io/projected/6d0058d8-63f2-4864-aa42-a9afbdfa721a-kube-api-access-7xvzs\") pod \"package-server-manager-789f6589d5-bw5th\" (UID: \"6d0058d8-63f2-4864-aa42-a9afbdfa721a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641138 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a7221db0-3f54-4e07-95ec-56c4fb95aecc-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-b5q8z\" (UID: \"a7221db0-3f54-4e07-95ec-56c4fb95aecc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641155 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fbc99729-cada-43b3-be3e-528a87db765e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641169 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1459a017-1dc3-4189-b81f-0f4dfcee04f4-srv-cert\") pod \"catalog-operator-68c6474976-tfgxd\" (UID: \"1459a017-1dc3-4189-b81f-0f4dfcee04f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641187 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/09b85eb4-ccad-4223-96a4-b53745e6f585-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-b6x96\" (UID: \"09b85eb4-ccad-4223-96a4-b53745e6f585\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641203 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-audit\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641220 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d11557e-c042-4dc0-bb61-3dfc27b309a1-config\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641235 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lcpx\" (UniqueName: \"kubernetes.io/projected/23ba3d67-97df-4a13-8603-4c0efa967852-kube-api-access-9lcpx\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641250 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7d11557e-c042-4dc0-bb61-3dfc27b309a1-etcd-service-ca\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641272 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdql7\" (UniqueName: \"kubernetes.io/projected/8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a-kube-api-access-vdql7\") pod \"multus-admission-controller-857f4d67dd-n26p9\" (UID: \"8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n26p9" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641286 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/10339d22-99de-4f28-98a2-f3549c0f4c13-encryption-config\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641304 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18e8bf6d-bea3-435c-9e4b-4d7a287f0c83-config\") pod \"machine-approver-56656f9798-nn48b\" (UID: \"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641319 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-n26p9\" (UID: \"8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n26p9" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641334 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fbc99729-cada-43b3-be3e-528a87db765e-audit-dir\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641351 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc5708d4-d312-48a6-931a-b09f0c1dd569-serving-cert\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641833 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svjqj\" (UniqueName: \"kubernetes.io/projected/ff9cb19e-ad39-4636-b302-57b5f39b0072-kube-api-access-svjqj\") pod \"route-controller-manager-6576b87f9c-hns6n\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641866 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/18e8bf6d-bea3-435c-9e4b-4d7a287f0c83-machine-approver-tls\") pod \"machine-approver-56656f9798-nn48b\" (UID: \"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641883 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/10339d22-99de-4f28-98a2-f3549c0f4c13-node-pullsecrets\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641898 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62b2562b-f25e-4276-bab1-0b366d60c231-trusted-ca\") pod \"ingress-operator-5b745b69d9-n9d98\" (UID: \"62b2562b-f25e-4276-bab1-0b366d60c231\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641915 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff9cb19e-ad39-4636-b302-57b5f39b0072-serving-cert\") pod \"route-controller-manager-6576b87f9c-hns6n\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641930 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r89m8\" (UniqueName: \"kubernetes.io/projected/1459a017-1dc3-4189-b81f-0f4dfcee04f4-kube-api-access-r89m8\") pod \"catalog-operator-68c6474976-tfgxd\" (UID: \"1459a017-1dc3-4189-b81f-0f4dfcee04f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641946 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4kxf\" (UniqueName: \"kubernetes.io/projected/382e1d8e-0916-487b-9f6c-2d946da409f1-kube-api-access-k4kxf\") pod \"marketplace-operator-79b997595-vkldx\" (UID: \"382e1d8e-0916-487b-9f6c-2d946da409f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641964 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bacbc855-fe99-4250-9da3-b64e95ac7c67-config-volume\") pod \"collect-profiles-29331900-b46lv\" (UID: \"bacbc855-fe99-4250-9da3-b64e95ac7c67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641980 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10339d22-99de-4f28-98a2-f3549c0f4c13-serving-cert\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.641996 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fbc99729-cada-43b3-be3e-528a87db765e-encryption-config\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642014 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrxkx\" (UniqueName: \"kubernetes.io/projected/7d11557e-c042-4dc0-bb61-3dfc27b309a1-kube-api-access-xrxkx\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642068 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/62b2562b-f25e-4276-bab1-0b366d60c231-metrics-tls\") pod \"ingress-operator-5b745b69d9-n9d98\" (UID: \"62b2562b-f25e-4276-bab1-0b366d60c231\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4vh5\" (UniqueName: \"kubernetes.io/projected/62b2562b-f25e-4276-bab1-0b366d60c231-kube-api-access-f4vh5\") pod \"ingress-operator-5b745b69d9-n9d98\" (UID: \"62b2562b-f25e-4276-bab1-0b366d60c231\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642105 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-config\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642124 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/23ba3d67-97df-4a13-8603-4c0efa967852-stats-auth\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642151 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fbc99729-cada-43b3-be3e-528a87db765e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642171 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/382e1d8e-0916-487b-9f6c-2d946da409f1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vkldx\" (UID: \"382e1d8e-0916-487b-9f6c-2d946da409f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642190 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18e8bf6d-bea3-435c-9e4b-4d7a287f0c83-auth-proxy-config\") pod \"machine-approver-56656f9798-nn48b\" (UID: \"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642210 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7d11557e-c042-4dc0-bb61-3dfc27b309a1-etcd-client\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642231 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae32f38d-23e4-402e-9475-9ea45ba692f8-config\") pod \"kube-apiserver-operator-766d6c64bb-tkvdh\" (UID: \"ae32f38d-23e4-402e-9475-9ea45ba692f8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642250 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9d7eea67-1ce4-4963-9a2d-577d18ce34e5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hxs46\" (UID: \"9d7eea67-1ce4-4963-9a2d-577d18ce34e5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642269 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fbc99729-cada-43b3-be3e-528a87db765e-etcd-client\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642300 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26t5q\" (UniqueName: \"kubernetes.io/projected/fbc99729-cada-43b3-be3e-528a87db765e-kube-api-access-26t5q\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642316 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff9cb19e-ad39-4636-b302-57b5f39b0072-config\") pod \"route-controller-manager-6576b87f9c-hns6n\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642331 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fbc99729-cada-43b3-be3e-528a87db765e-audit-policies\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642347 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc5708d4-d312-48a6-931a-b09f0c1dd569-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642378 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642393 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae32f38d-23e4-402e-9475-9ea45ba692f8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-tkvdh\" (UID: \"ae32f38d-23e4-402e-9475-9ea45ba692f8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642422 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hls6m\" (UniqueName: \"kubernetes.io/projected/09b85eb4-ccad-4223-96a4-b53745e6f585-kube-api-access-hls6m\") pod \"cluster-samples-operator-665b6dd947-b6x96\" (UID: \"09b85eb4-ccad-4223-96a4-b53745e6f585\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642441 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc5708d4-d312-48a6-931a-b09f0c1dd569-config\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642458 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/10339d22-99de-4f28-98a2-f3549c0f4c13-audit-dir\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642473 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/382e1d8e-0916-487b-9f6c-2d946da409f1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vkldx\" (UID: \"382e1d8e-0916-487b-9f6c-2d946da409f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642491 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvxs2\" (UniqueName: \"kubernetes.io/projected/a7221db0-3f54-4e07-95ec-56c4fb95aecc-kube-api-access-dvxs2\") pod \"machine-api-operator-5694c8668f-b5q8z\" (UID: \"a7221db0-3f54-4e07-95ec-56c4fb95aecc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642509 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9d7eea67-1ce4-4963-9a2d-577d18ce34e5-srv-cert\") pod \"olm-operator-6b444d44fb-hxs46\" (UID: \"9d7eea67-1ce4-4963-9a2d-577d18ce34e5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642525 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/23ba3d67-97df-4a13-8603-4c0efa967852-service-ca-bundle\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642541 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7d11557e-c042-4dc0-bb61-3dfc27b309a1-etcd-ca\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642558 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/800b4855-f900-4968-9a86-fad43cec8fa6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-99rp7\" (UID: \"800b4855-f900-4968-9a86-fad43cec8fa6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642593 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtjgm\" (UniqueName: \"kubernetes.io/projected/10339d22-99de-4f28-98a2-f3549c0f4c13-kube-api-access-xtjgm\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642609 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m6qb\" (UniqueName: \"kubernetes.io/projected/431d0dc1-8b42-47fb-9a90-71e957e3bddc-kube-api-access-7m6qb\") pod \"migrator-59844c95c7-jgg9g\" (UID: \"431d0dc1-8b42-47fb-9a90-71e957e3bddc\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jgg9g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642627 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1459a017-1dc3-4189-b81f-0f4dfcee04f4-profile-collector-cert\") pod \"catalog-operator-68c6474976-tfgxd\" (UID: \"1459a017-1dc3-4189-b81f-0f4dfcee04f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642649 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae32f38d-23e4-402e-9475-9ea45ba692f8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-tkvdh\" (UID: \"ae32f38d-23e4-402e-9475-9ea45ba692f8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642665 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44j7h\" (UniqueName: \"kubernetes.io/projected/9d7eea67-1ce4-4963-9a2d-577d18ce34e5-kube-api-access-44j7h\") pod \"olm-operator-6b444d44fb-hxs46\" (UID: \"9d7eea67-1ce4-4963-9a2d-577d18ce34e5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642689 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc63fe57-540e-4c29-8fc6-155d220f2e9e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m4wbt\" (UID: \"fc63fe57-540e-4c29-8fc6-155d220f2e9e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642705 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbc99729-cada-43b3-be3e-528a87db765e-serving-cert\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642720 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/23ba3d67-97df-4a13-8603-4c0efa967852-metrics-certs\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642735 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/10339d22-99de-4f28-98a2-f3549c0f4c13-etcd-client\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q76n\" (UniqueName: \"kubernetes.io/projected/81cf0144-e970-4fe2-8663-15b006dc1b91-kube-api-access-2q76n\") pod \"control-plane-machine-set-operator-78cbb6b69f-5nnfc\" (UID: \"81cf0144-e970-4fe2-8663-15b006dc1b91\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642771 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05957de0-ed69-46e2-aab9-d146cfcbbb90-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-54nth\" (UID: \"05957de0-ed69-46e2-aab9-d146cfcbbb90\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642787 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/800b4855-f900-4968-9a86-fad43cec8fa6-config\") pod \"kube-controller-manager-operator-78b949d7b-99rp7\" (UID: \"800b4855-f900-4968-9a86-fad43cec8fa6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642801 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-etcd-serving-ca\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642818 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d0058d8-63f2-4864-aa42-a9afbdfa721a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bw5th\" (UID: \"6d0058d8-63f2-4864-aa42-a9afbdfa721a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642836 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc63fe57-540e-4c29-8fc6-155d220f2e9e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m4wbt\" (UID: \"fc63fe57-540e-4c29-8fc6-155d220f2e9e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642853 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/62b2562b-f25e-4276-bab1-0b366d60c231-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n9d98\" (UID: \"62b2562b-f25e-4276-bab1-0b366d60c231\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642869 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d11557e-c042-4dc0-bb61-3dfc27b309a1-serving-cert\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642884 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhrvh\" (UniqueName: \"kubernetes.io/projected/18e8bf6d-bea3-435c-9e4b-4d7a287f0c83-kube-api-access-zhrvh\") pod \"machine-approver-56656f9798-nn48b\" (UID: \"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642899 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05957de0-ed69-46e2-aab9-d146cfcbbb90-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-54nth\" (UID: \"05957de0-ed69-46e2-aab9-d146cfcbbb90\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642915 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc63fe57-540e-4c29-8fc6-155d220f2e9e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m4wbt\" (UID: \"fc63fe57-540e-4c29-8fc6-155d220f2e9e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642931 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bacbc855-fe99-4250-9da3-b64e95ac7c67-secret-volume\") pod \"collect-profiles-29331900-b46lv\" (UID: \"bacbc855-fe99-4250-9da3-b64e95ac7c67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642950 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rzws\" (UniqueName: \"kubernetes.io/projected/dc5708d4-d312-48a6-931a-b09f0c1dd569-kube-api-access-9rzws\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642968 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc5708d4-d312-48a6-931a-b09f0c1dd569-service-ca-bundle\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.642983 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff9cb19e-ad39-4636-b302-57b5f39b0072-client-ca\") pod \"route-controller-manager-6576b87f9c-hns6n\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.643429 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zc56g"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.643496 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x97sx"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.644527 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff9cb19e-ad39-4636-b302-57b5f39b0072-client-ca\") pod \"route-controller-manager-6576b87f9c-hns6n\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.647341 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dr8c5"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.647389 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-f8825"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.647401 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-jgg9g"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.648174 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fbc99729-cada-43b3-be3e-528a87db765e-audit-policies\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.649225 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fbc99729-cada-43b3-be3e-528a87db765e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.650787 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fbc99729-cada-43b3-be3e-528a87db765e-audit-dir\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.651552 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff9cb19e-ad39-4636-b302-57b5f39b0072-config\") pod \"route-controller-manager-6576b87f9c-hns6n\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.652352 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc5708d4-d312-48a6-931a-b09f0c1dd569-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.653806 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7221db0-3f54-4e07-95ec-56c4fb95aecc-config\") pod \"machine-api-operator-5694c8668f-b5q8z\" (UID: \"a7221db0-3f54-4e07-95ec-56c4fb95aecc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.654412 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.654967 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hflmk"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.655063 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9knh2"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.655111 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/10339d22-99de-4f28-98a2-f3549c0f4c13-node-pullsecrets\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.654931 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.655936 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc5708d4-d312-48a6-931a-b09f0c1dd569-config\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.656135 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/10339d22-99de-4f28-98a2-f3549c0f4c13-audit-dir\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.656352 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62b2562b-f25e-4276-bab1-0b366d60c231-trusted-ca\") pod \"ingress-operator-5b745b69d9-n9d98\" (UID: \"62b2562b-f25e-4276-bab1-0b366d60c231\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.659047 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc5708d4-d312-48a6-931a-b09f0c1dd569-serving-cert\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.659659 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bacbc855-fe99-4250-9da3-b64e95ac7c67-config-volume\") pod \"collect-profiles-29331900-b46lv\" (UID: \"bacbc855-fe99-4250-9da3-b64e95ac7c67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.660079 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/23ba3d67-97df-4a13-8603-4c0efa967852-metrics-certs\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.661698 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-audit\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.662622 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/18e8bf6d-bea3-435c-9e4b-4d7a287f0c83-machine-approver-tls\") pod \"machine-approver-56656f9798-nn48b\" (UID: \"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.663502 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-image-import-ca\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.664563 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-etcd-serving-ca\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.664807 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc63fe57-540e-4c29-8fc6-155d220f2e9e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m4wbt\" (UID: \"fc63fe57-540e-4c29-8fc6-155d220f2e9e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.665037 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-config\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.665342 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.665395 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5kssv"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.665559 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/800b4855-f900-4968-9a86-fad43cec8fa6-config\") pod \"kube-controller-manager-operator-78b949d7b-99rp7\" (UID: \"800b4855-f900-4968-9a86-fad43cec8fa6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.666038 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05957de0-ed69-46e2-aab9-d146cfcbbb90-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-54nth\" (UID: \"05957de0-ed69-46e2-aab9-d146cfcbbb90\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.667143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18e8bf6d-bea3-435c-9e4b-4d7a287f0c83-config\") pod \"machine-approver-56656f9798-nn48b\" (UID: \"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.667474 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fbc99729-cada-43b3-be3e-528a87db765e-encryption-config\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.667677 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a7221db0-3f54-4e07-95ec-56c4fb95aecc-images\") pod \"machine-api-operator-5694c8668f-b5q8z\" (UID: \"a7221db0-3f54-4e07-95ec-56c4fb95aecc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.667923 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1459a017-1dc3-4189-b81f-0f4dfcee04f4-srv-cert\") pod \"catalog-operator-68c6474976-tfgxd\" (UID: \"1459a017-1dc3-4189-b81f-0f4dfcee04f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.667972 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05957de0-ed69-46e2-aab9-d146cfcbbb90-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-54nth\" (UID: \"05957de0-ed69-46e2-aab9-d146cfcbbb90\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.668020 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.668046 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-tbc7j"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.668096 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/10339d22-99de-4f28-98a2-f3549c0f4c13-encryption-config\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.668827 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc63fe57-540e-4c29-8fc6-155d220f2e9e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m4wbt\" (UID: \"fc63fe57-540e-4c29-8fc6-155d220f2e9e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.669230 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fbc99729-cada-43b3-be3e-528a87db765e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.669486 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc5708d4-d312-48a6-931a-b09f0c1dd569-service-ca-bundle\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.669840 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/23ba3d67-97df-4a13-8603-4c0efa967852-service-ca-bundle\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.670249 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18e8bf6d-bea3-435c-9e4b-4d7a287f0c83-auth-proxy-config\") pod \"machine-approver-56656f9798-nn48b\" (UID: \"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.670695 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/62b2562b-f25e-4276-bab1-0b366d60c231-metrics-tls\") pod \"ingress-operator-5b745b69d9-n9d98\" (UID: \"62b2562b-f25e-4276-bab1-0b366d60c231\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.671944 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/10339d22-99de-4f28-98a2-f3549c0f4c13-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.672408 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a7221db0-3f54-4e07-95ec-56c4fb95aecc-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-b5q8z\" (UID: \"a7221db0-3f54-4e07-95ec-56c4fb95aecc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.673030 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbc99729-cada-43b3-be3e-528a87db765e-serving-cert\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.673663 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff9cb19e-ad39-4636-b302-57b5f39b0072-serving-cert\") pod \"route-controller-manager-6576b87f9c-hns6n\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.674647 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-rds4c"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.676040 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rds4c"] Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.676134 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.677675 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/23ba3d67-97df-4a13-8603-4c0efa967852-default-certificate\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.679396 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/81cf0144-e970-4fe2-8663-15b006dc1b91-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5nnfc\" (UID: \"81cf0144-e970-4fe2-8663-15b006dc1b91\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.680589 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/10339d22-99de-4f28-98a2-f3549c0f4c13-etcd-client\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.682383 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fbc99729-cada-43b3-be3e-528a87db765e-etcd-client\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.682559 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/23ba3d67-97df-4a13-8603-4c0efa967852-stats-auth\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.682395 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bacbc855-fe99-4250-9da3-b64e95ac7c67-secret-volume\") pod \"collect-profiles-29331900-b46lv\" (UID: \"bacbc855-fe99-4250-9da3-b64e95ac7c67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.682673 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10339d22-99de-4f28-98a2-f3549c0f4c13-serving-cert\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.682912 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-n26p9\" (UID: \"8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n26p9" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.683662 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.685797 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.686220 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/800b4855-f900-4968-9a86-fad43cec8fa6-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-99rp7\" (UID: \"800b4855-f900-4968-9a86-fad43cec8fa6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.694040 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1459a017-1dc3-4189-b81f-0f4dfcee04f4-profile-collector-cert\") pod \"catalog-operator-68c6474976-tfgxd\" (UID: \"1459a017-1dc3-4189-b81f-0f4dfcee04f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.704003 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.737188 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743435 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d11557e-c042-4dc0-bb61-3dfc27b309a1-serving-cert\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743481 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xvzs\" (UniqueName: \"kubernetes.io/projected/6d0058d8-63f2-4864-aa42-a9afbdfa721a-kube-api-access-7xvzs\") pod \"package-server-manager-789f6589d5-bw5th\" (UID: \"6d0058d8-63f2-4864-aa42-a9afbdfa721a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743499 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/09b85eb4-ccad-4223-96a4-b53745e6f585-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-b6x96\" (UID: \"09b85eb4-ccad-4223-96a4-b53745e6f585\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743515 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d11557e-c042-4dc0-bb61-3dfc27b309a1-config\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743541 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7d11557e-c042-4dc0-bb61-3dfc27b309a1-etcd-service-ca\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743577 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4kxf\" (UniqueName: \"kubernetes.io/projected/382e1d8e-0916-487b-9f6c-2d946da409f1-kube-api-access-k4kxf\") pod \"marketplace-operator-79b997595-vkldx\" (UID: \"382e1d8e-0916-487b-9f6c-2d946da409f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743594 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrxkx\" (UniqueName: \"kubernetes.io/projected/7d11557e-c042-4dc0-bb61-3dfc27b309a1-kube-api-access-xrxkx\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743620 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/382e1d8e-0916-487b-9f6c-2d946da409f1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vkldx\" (UID: \"382e1d8e-0916-487b-9f6c-2d946da409f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743634 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7d11557e-c042-4dc0-bb61-3dfc27b309a1-etcd-client\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743651 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae32f38d-23e4-402e-9475-9ea45ba692f8-config\") pod \"kube-apiserver-operator-766d6c64bb-tkvdh\" (UID: \"ae32f38d-23e4-402e-9475-9ea45ba692f8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743719 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9d7eea67-1ce4-4963-9a2d-577d18ce34e5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hxs46\" (UID: \"9d7eea67-1ce4-4963-9a2d-577d18ce34e5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743753 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae32f38d-23e4-402e-9475-9ea45ba692f8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-tkvdh\" (UID: \"ae32f38d-23e4-402e-9475-9ea45ba692f8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743784 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hls6m\" (UniqueName: \"kubernetes.io/projected/09b85eb4-ccad-4223-96a4-b53745e6f585-kube-api-access-hls6m\") pod \"cluster-samples-operator-665b6dd947-b6x96\" (UID: \"09b85eb4-ccad-4223-96a4-b53745e6f585\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743808 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/382e1d8e-0916-487b-9f6c-2d946da409f1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vkldx\" (UID: \"382e1d8e-0916-487b-9f6c-2d946da409f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743825 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9d7eea67-1ce4-4963-9a2d-577d18ce34e5-srv-cert\") pod \"olm-operator-6b444d44fb-hxs46\" (UID: \"9d7eea67-1ce4-4963-9a2d-577d18ce34e5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743844 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7d11557e-c042-4dc0-bb61-3dfc27b309a1-etcd-ca\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743873 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m6qb\" (UniqueName: \"kubernetes.io/projected/431d0dc1-8b42-47fb-9a90-71e957e3bddc-kube-api-access-7m6qb\") pod \"migrator-59844c95c7-jgg9g\" (UID: \"431d0dc1-8b42-47fb-9a90-71e957e3bddc\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jgg9g" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.743891 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae32f38d-23e4-402e-9475-9ea45ba692f8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-tkvdh\" (UID: \"ae32f38d-23e4-402e-9475-9ea45ba692f8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.744017 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44j7h\" (UniqueName: \"kubernetes.io/projected/9d7eea67-1ce4-4963-9a2d-577d18ce34e5-kube-api-access-44j7h\") pod \"olm-operator-6b444d44fb-hxs46\" (UID: \"9d7eea67-1ce4-4963-9a2d-577d18ce34e5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.744102 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d0058d8-63f2-4864-aa42-a9afbdfa721a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bw5th\" (UID: \"6d0058d8-63f2-4864-aa42-a9afbdfa721a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.746251 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/382e1d8e-0916-487b-9f6c-2d946da409f1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vkldx\" (UID: \"382e1d8e-0916-487b-9f6c-2d946da409f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.748919 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/382e1d8e-0916-487b-9f6c-2d946da409f1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vkldx\" (UID: \"382e1d8e-0916-487b-9f6c-2d946da409f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.749349 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9d7eea67-1ce4-4963-9a2d-577d18ce34e5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hxs46\" (UID: \"9d7eea67-1ce4-4963-9a2d-577d18ce34e5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.749799 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d0058d8-63f2-4864-aa42-a9afbdfa721a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bw5th\" (UID: \"6d0058d8-63f2-4864-aa42-a9afbdfa721a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.755376 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.767606 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.784338 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.804123 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.824414 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.844006 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.864166 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.888895 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.904611 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.924034 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.944385 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.950508 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/09b85eb4-ccad-4223-96a4-b53745e6f585-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-b6x96\" (UID: \"09b85eb4-ccad-4223-96a4-b53745e6f585\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96" Oct 08 09:14:03 crc kubenswrapper[4744]: I1008 09:14:03.964131 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.005053 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.005155 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.025015 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.044570 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.064153 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.084621 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.086558 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7d11557e-c042-4dc0-bb61-3dfc27b309a1-etcd-service-ca\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.103990 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.124701 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.145292 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.165052 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.184686 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.204917 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.224856 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.244905 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.264318 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.284126 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.304327 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.324688 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.330062 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d11557e-c042-4dc0-bb61-3dfc27b309a1-serving-cert\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.344575 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.347934 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7d11557e-c042-4dc0-bb61-3dfc27b309a1-etcd-client\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.364995 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.384510 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.404911 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.407742 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d11557e-c042-4dc0-bb61-3dfc27b309a1-config\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.424527 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.425983 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7d11557e-c042-4dc0-bb61-3dfc27b309a1-etcd-ca\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.465085 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.484997 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.504309 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.542514 4744 request.go:700] Waited for 1.007606611s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.544708 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.564665 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.584927 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.604890 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.606903 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae32f38d-23e4-402e-9475-9ea45ba692f8-config\") pod \"kube-apiserver-operator-766d6c64bb-tkvdh\" (UID: \"ae32f38d-23e4-402e-9475-9ea45ba692f8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.624861 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.629834 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9d7eea67-1ce4-4963-9a2d-577d18ce34e5-srv-cert\") pod \"olm-operator-6b444d44fb-hxs46\" (UID: \"9d7eea67-1ce4-4963-9a2d-577d18ce34e5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.645310 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.664353 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.670408 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae32f38d-23e4-402e-9475-9ea45ba692f8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-tkvdh\" (UID: \"ae32f38d-23e4-402e-9475-9ea45ba692f8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.685329 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.704341 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.724346 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.745000 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.765682 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.785148 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.805275 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.824696 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.844657 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.865534 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.884700 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.904525 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.924956 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.945130 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.964803 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 08 09:14:04 crc kubenswrapper[4744]: I1008 09:14:04.985058 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.005046 4744 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.025579 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.044701 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.066708 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.085826 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.104794 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.125747 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.144117 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.164152 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.204987 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.205590 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.223700 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.245982 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.263721 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.284869 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.304422 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.324912 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.345089 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.378533 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fc63fe57-540e-4c29-8fc6-155d220f2e9e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-m4wbt\" (UID: \"fc63fe57-540e-4c29-8fc6-155d220f2e9e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.388551 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.408448 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26t5q\" (UniqueName: \"kubernetes.io/projected/fbc99729-cada-43b3-be3e-528a87db765e-kube-api-access-26t5q\") pod \"apiserver-7bbb656c7d-7ll8m\" (UID: \"fbc99729-cada-43b3-be3e-528a87db765e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.426216 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svjqj\" (UniqueName: \"kubernetes.io/projected/ff9cb19e-ad39-4636-b302-57b5f39b0072-kube-api-access-svjqj\") pod \"route-controller-manager-6576b87f9c-hns6n\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.439019 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvxs2\" (UniqueName: \"kubernetes.io/projected/a7221db0-3f54-4e07-95ec-56c4fb95aecc-kube-api-access-dvxs2\") pod \"machine-api-operator-5694c8668f-b5q8z\" (UID: \"a7221db0-3f54-4e07-95ec-56c4fb95aecc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.462796 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r89m8\" (UniqueName: \"kubernetes.io/projected/1459a017-1dc3-4189-b81f-0f4dfcee04f4-kube-api-access-r89m8\") pod \"catalog-operator-68c6474976-tfgxd\" (UID: \"1459a017-1dc3-4189-b81f-0f4dfcee04f4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.481613 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.484296 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lcpx\" (UniqueName: \"kubernetes.io/projected/23ba3d67-97df-4a13-8603-4c0efa967852-kube-api-access-9lcpx\") pod \"router-default-5444994796-5j25g\" (UID: \"23ba3d67-97df-4a13-8603-4c0efa967852\") " pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.504114 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdql7\" (UniqueName: \"kubernetes.io/projected/8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a-kube-api-access-vdql7\") pod \"multus-admission-controller-857f4d67dd-n26p9\" (UID: \"8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-n26p9" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.505673 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.524439 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdnxn\" (UniqueName: \"kubernetes.io/projected/bacbc855-fe99-4250-9da3-b64e95ac7c67-kube-api-access-wdnxn\") pod \"collect-profiles-29331900-b46lv\" (UID: \"bacbc855-fe99-4250-9da3-b64e95ac7c67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.542863 4744 request.go:700] Waited for 1.878745779s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-machine-approver/serviceaccounts/machine-approver-sa/token Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.549146 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q76n\" (UniqueName: \"kubernetes.io/projected/81cf0144-e970-4fe2-8663-15b006dc1b91-kube-api-access-2q76n\") pod \"control-plane-machine-set-operator-78cbb6b69f-5nnfc\" (UID: \"81cf0144-e970-4fe2-8663-15b006dc1b91\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.559932 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.570958 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhrvh\" (UniqueName: \"kubernetes.io/projected/18e8bf6d-bea3-435c-9e4b-4d7a287f0c83-kube-api-access-zhrvh\") pod \"machine-approver-56656f9798-nn48b\" (UID: \"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.585073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/62b2562b-f25e-4276-bab1-0b366d60c231-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n9d98\" (UID: \"62b2562b-f25e-4276-bab1-0b366d60c231\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.606847 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4vh5\" (UniqueName: \"kubernetes.io/projected/62b2562b-f25e-4276-bab1-0b366d60c231-kube-api-access-f4vh5\") pod \"ingress-operator-5b745b69d9-n9d98\" (UID: \"62b2562b-f25e-4276-bab1-0b366d60c231\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.615579 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.627129 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p59xk\" (UniqueName: \"kubernetes.io/projected/05957de0-ed69-46e2-aab9-d146cfcbbb90-kube-api-access-p59xk\") pod \"kube-storage-version-migrator-operator-b67b599dd-54nth\" (UID: \"05957de0-ed69-46e2-aab9-d146cfcbbb90\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.638073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rzws\" (UniqueName: \"kubernetes.io/projected/dc5708d4-d312-48a6-931a-b09f0c1dd569-kube-api-access-9rzws\") pod \"authentication-operator-69f744f599-fl2qz\" (UID: \"dc5708d4-d312-48a6-931a-b09f0c1dd569\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.660100 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/800b4855-f900-4968-9a86-fad43cec8fa6-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-99rp7\" (UID: \"800b4855-f900-4968-9a86-fad43cec8fa6\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.662090 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.677335 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.677820 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtjgm\" (UniqueName: \"kubernetes.io/projected/10339d22-99de-4f28-98a2-f3549c0f4c13-kube-api-access-xtjgm\") pod \"apiserver-76f77b778f-fjhhp\" (UID: \"10339d22-99de-4f28-98a2-f3549c0f4c13\") " pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.684960 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 08 09:14:05 crc kubenswrapper[4744]: W1008 09:14:05.685639 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23ba3d67_97df_4a13_8603_4c0efa967852.slice/crio-96b2b0b3845e735bf50e16faf74c8037da481af0e394cef13fc7664bbf03b220 WatchSource:0}: Error finding container 96b2b0b3845e735bf50e16faf74c8037da481af0e394cef13fc7664bbf03b220: Status 404 returned error can't find the container with id 96b2b0b3845e735bf50e16faf74c8037da481af0e394cef13fc7664bbf03b220 Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.695235 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.707766 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.719193 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.724557 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.730766 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-n26p9" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.748836 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.750917 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n"] Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.778345 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt"] Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.785038 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m6qb\" (UniqueName: \"kubernetes.io/projected/431d0dc1-8b42-47fb-9a90-71e957e3bddc-kube-api-access-7m6qb\") pod \"migrator-59844c95c7-jgg9g\" (UID: \"431d0dc1-8b42-47fb-9a90-71e957e3bddc\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jgg9g" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.785065 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xvzs\" (UniqueName: \"kubernetes.io/projected/6d0058d8-63f2-4864-aa42-a9afbdfa721a-kube-api-access-7xvzs\") pod \"package-server-manager-789f6589d5-bw5th\" (UID: \"6d0058d8-63f2-4864-aa42-a9afbdfa721a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.801046 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hls6m\" (UniqueName: \"kubernetes.io/projected/09b85eb4-ccad-4223-96a4-b53745e6f585-kube-api-access-hls6m\") pod \"cluster-samples-operator-665b6dd947-b6x96\" (UID: \"09b85eb4-ccad-4223-96a4-b53745e6f585\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.808488 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.815114 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.827863 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4kxf\" (UniqueName: \"kubernetes.io/projected/382e1d8e-0916-487b-9f6c-2d946da409f1-kube-api-access-k4kxf\") pod \"marketplace-operator-79b997595-vkldx\" (UID: \"382e1d8e-0916-487b-9f6c-2d946da409f1\") " pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.830805 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.835493 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.841836 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrxkx\" (UniqueName: \"kubernetes.io/projected/7d11557e-c042-4dc0-bb61-3dfc27b309a1-kube-api-access-xrxkx\") pod \"etcd-operator-b45778765-cf7sw\" (UID: \"7d11557e-c042-4dc0-bb61-3dfc27b309a1\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.842024 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.842290 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.870301 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.878836 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae32f38d-23e4-402e-9475-9ea45ba692f8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-tkvdh\" (UID: \"ae32f38d-23e4-402e-9475-9ea45ba692f8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.881256 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44j7h\" (UniqueName: \"kubernetes.io/projected/9d7eea67-1ce4-4963-9a2d-577d18ce34e5-kube-api-access-44j7h\") pod \"olm-operator-6b444d44fb-hxs46\" (UID: \"9d7eea67-1ce4-4963-9a2d-577d18ce34e5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.885925 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jgg9g" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.920298 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.927514 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.933291 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.984339 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-b5q8z"] Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997312 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997347 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcn5g\" (UniqueName: \"kubernetes.io/projected/d17e7154-7a80-4607-84d7-016ff6eb07c5-kube-api-access-dcn5g\") pod \"cluster-image-registry-operator-dc59b4c8b-l2l6r\" (UID: \"d17e7154-7a80-4607-84d7-016ff6eb07c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997389 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b2217f7-f478-4833-b301-097e7c1c07c9-trusted-ca\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997416 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn44b\" (UniqueName: \"kubernetes.io/projected/41d15067-9885-4940-80f8-fbd63458c326-kube-api-access-wn44b\") pod \"service-ca-operator-777779d784-hflmk\" (UID: \"41d15067-9885-4940-80f8-fbd63458c326\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997450 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997483 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqc5c\" (UniqueName: \"kubernetes.io/projected/70231056-c93b-47e7-a2f0-a44329e6f033-kube-api-access-vqc5c\") pod \"openshift-apiserver-operator-796bbdcf4f-xflpt\" (UID: \"70231056-c93b-47e7-a2f0-a44329e6f033\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997510 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v2d5\" (UniqueName: \"kubernetes.io/projected/3f64391a-1d93-4e10-9a9a-0b878bde535b-kube-api-access-5v2d5\") pod \"console-operator-58897d9998-pjz2t\" (UID: \"3f64391a-1d93-4e10-9a9a-0b878bde535b\") " pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997558 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d17e7154-7a80-4607-84d7-016ff6eb07c5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-l2l6r\" (UID: \"d17e7154-7a80-4607-84d7-016ff6eb07c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997609 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997662 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997703 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlf55\" (UniqueName: \"kubernetes.io/projected/216b2020-4266-4895-85cb-f0939118b7a8-kube-api-access-mlf55\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997734 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m76cx\" (UniqueName: \"kubernetes.io/projected/14215523-611c-4c7d-a2a9-f4dbe478b0e0-kube-api-access-m76cx\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997805 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1b2217f7-f478-4833-b301-097e7c1c07c9-registry-certificates\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997825 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-config\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997856 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zklvg\" (UID: \"7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997877 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3f64391a-1d93-4e10-9a9a-0b878bde535b-trusted-ca\") pod \"console-operator-58897d9998-pjz2t\" (UID: \"3f64391a-1d93-4e10-9a9a-0b878bde535b\") " pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997896 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997914 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997933 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4bdr\" (UniqueName: \"kubernetes.io/projected/4a4443bb-3033-4719-b515-982c9f3f776a-kube-api-access-l4bdr\") pod \"machine-config-operator-74547568cd-f8825\" (UID: \"4a4443bb-3033-4719-b515-982c9f3f776a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997972 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f-serving-cert\") pod \"openshift-config-operator-7777fb866f-zklvg\" (UID: \"7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.997990 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41d15067-9885-4940-80f8-fbd63458c326-config\") pod \"service-ca-operator-777779d784-hflmk\" (UID: \"41d15067-9885-4940-80f8-fbd63458c326\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.998021 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4a4443bb-3033-4719-b515-982c9f3f776a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-f8825\" (UID: \"4a4443bb-3033-4719-b515-982c9f3f776a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.998111 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d5481818-bc94-4c08-9bfe-444c1ed82e1a-metrics-tls\") pod \"dns-operator-744455d44c-dr8c5\" (UID: \"d5481818-bc94-4c08-9bfe-444c1ed82e1a\") " pod="openshift-dns-operator/dns-operator-744455d44c-dr8c5" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.998141 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70231056-c93b-47e7-a2f0-a44329e6f033-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xflpt\" (UID: \"70231056-c93b-47e7-a2f0-a44329e6f033\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.998182 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-registry-tls\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.998201 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-audit-policies\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.998220 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41d15067-9885-4940-80f8-fbd63458c326-serving-cert\") pod \"service-ca-operator-777779d784-hflmk\" (UID: \"41d15067-9885-4940-80f8-fbd63458c326\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999530 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1b2217f7-f478-4833-b301-097e7c1c07c9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14215523-611c-4c7d-a2a9-f4dbe478b0e0-audit-dir\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999598 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70231056-c93b-47e7-a2f0-a44329e6f033-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xflpt\" (UID: \"70231056-c93b-47e7-a2f0-a44329e6f033\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999622 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d17e7154-7a80-4607-84d7-016ff6eb07c5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-l2l6r\" (UID: \"d17e7154-7a80-4607-84d7-016ff6eb07c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999645 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/216b2020-4266-4895-85cb-f0939118b7a8-serving-cert\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999677 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999723 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4a4443bb-3033-4719-b515-982c9f3f776a-proxy-tls\") pod \"machine-config-operator-74547568cd-f8825\" (UID: \"4a4443bb-3033-4719-b515-982c9f3f776a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999739 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999762 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d17e7154-7a80-4607-84d7-016ff6eb07c5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-l2l6r\" (UID: \"d17e7154-7a80-4607-84d7-016ff6eb07c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999779 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999794 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn9jq\" (UniqueName: \"kubernetes.io/projected/3dd5759b-8411-48b6-a537-422f23c7d00d-kube-api-access-dn9jq\") pod \"service-ca-9c57cc56f-zc56g\" (UID: \"3dd5759b-8411-48b6-a537-422f23c7d00d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999818 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4a4443bb-3033-4719-b515-982c9f3f776a-images\") pod \"machine-config-operator-74547568cd-f8825\" (UID: \"4a4443bb-3033-4719-b515-982c9f3f776a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999856 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3dd5759b-8411-48b6-a537-422f23c7d00d-signing-key\") pod \"service-ca-9c57cc56f-zc56g\" (UID: \"3dd5759b-8411-48b6-a537-422f23c7d00d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999875 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3dd5759b-8411-48b6-a537-422f23c7d00d-signing-cabundle\") pod \"service-ca-9c57cc56f-zc56g\" (UID: \"3dd5759b-8411-48b6-a537-422f23c7d00d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999902 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1b2217f7-f478-4833-b301-097e7c1c07c9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999923 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-client-ca\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999936 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f64391a-1d93-4e10-9a9a-0b878bde535b-config\") pod \"console-operator-58897d9998-pjz2t\" (UID: \"3f64391a-1d93-4e10-9a9a-0b878bde535b\") " pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:05 crc kubenswrapper[4744]: I1008 09:14:05.999959 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:05.999974 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:05.999990 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t88tc\" (UniqueName: \"kubernetes.io/projected/05dd98d8-f557-49ea-84ce-2c23b9de7220-kube-api-access-t88tc\") pod \"downloads-7954f5f757-8whxk\" (UID: \"05dd98d8-f557-49ea-84ce-2c23b9de7220\") " pod="openshift-console/downloads-7954f5f757-8whxk" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.000007 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-bound-sa-token\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.000021 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.000045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2hpx\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-kube-api-access-s2hpx\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.000059 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc2fp\" (UniqueName: \"kubernetes.io/projected/7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f-kube-api-access-wc2fp\") pod \"openshift-config-operator-7777fb866f-zklvg\" (UID: \"7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.000094 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f64391a-1d93-4e10-9a9a-0b878bde535b-serving-cert\") pod \"console-operator-58897d9998-pjz2t\" (UID: \"3f64391a-1d93-4e10-9a9a-0b878bde535b\") " pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.000138 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7cvp\" (UniqueName: \"kubernetes.io/projected/d5481818-bc94-4c08-9bfe-444c1ed82e1a-kube-api-access-c7cvp\") pod \"dns-operator-744455d44c-dr8c5\" (UID: \"d5481818-bc94-4c08-9bfe-444c1ed82e1a\") " pod="openshift-dns-operator/dns-operator-744455d44c-dr8c5" Oct 08 09:14:06 crc kubenswrapper[4744]: E1008 09:14:06.003863 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:06.503850627 +0000 UTC m=+141.751495866 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.032566 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth"] Oct 08 09:14:06 crc kubenswrapper[4744]: W1008 09:14:06.046577 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7221db0_3f54_4e07_95ec_56c4fb95aecc.slice/crio-c8a30d46737c0817396209e145c0490b715f28324a11718802e39d516a98bf7e WatchSource:0}: Error finding container c8a30d46737c0817396209e145c0490b715f28324a11718802e39d516a98bf7e: Status 404 returned error can't find the container with id c8a30d46737c0817396209e145c0490b715f28324a11718802e39d516a98bf7e Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.100818 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m"] Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101266 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101495 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7cvp\" (UniqueName: \"kubernetes.io/projected/d5481818-bc94-4c08-9bfe-444c1ed82e1a-kube-api-access-c7cvp\") pod \"dns-operator-744455d44c-dr8c5\" (UID: \"d5481818-bc94-4c08-9bfe-444c1ed82e1a\") " pod="openshift-dns-operator/dns-operator-744455d44c-dr8c5" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101518 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f64391a-1d93-4e10-9a9a-0b878bde535b-serving-cert\") pod \"console-operator-58897d9998-pjz2t\" (UID: \"3f64391a-1d93-4e10-9a9a-0b878bde535b\") " pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101541 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-oauth-serving-cert\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb805699-c562-4a60-a45a-67dbbba49779-console-oauth-config\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101581 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-console-config\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101601 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/275d1220-b88e-4ff4-a07d-332cb73bb4fa-cert\") pod \"ingress-canary-tbc7j\" (UID: \"275d1220-b88e-4ff4-a07d-332cb73bb4fa\") " pod="openshift-ingress-canary/ingress-canary-tbc7j" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101620 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b2217f7-f478-4833-b301-097e7c1c07c9-trusted-ca\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101639 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101658 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcn5g\" (UniqueName: \"kubernetes.io/projected/d17e7154-7a80-4607-84d7-016ff6eb07c5-kube-api-access-dcn5g\") pod \"cluster-image-registry-operator-dc59b4c8b-l2l6r\" (UID: \"d17e7154-7a80-4607-84d7-016ff6eb07c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101678 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvhfr\" (UniqueName: \"kubernetes.io/projected/da22505d-c178-4739-ac34-0d6750469f7b-kube-api-access-dvhfr\") pod \"machine-config-controller-84d6567774-2p8wm\" (UID: \"da22505d-c178-4739-ac34-0d6750469f7b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101700 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn44b\" (UniqueName: \"kubernetes.io/projected/41d15067-9885-4940-80f8-fbd63458c326-kube-api-access-wn44b\") pod \"service-ca-operator-777779d784-hflmk\" (UID: \"41d15067-9885-4940-80f8-fbd63458c326\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101718 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-socket-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.101741 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:06 crc kubenswrapper[4744]: E1008 09:14:06.103634 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:06.603592543 +0000 UTC m=+141.851237792 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.106710 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqc5c\" (UniqueName: \"kubernetes.io/projected/70231056-c93b-47e7-a2f0-a44329e6f033-kube-api-access-vqc5c\") pod \"openshift-apiserver-operator-796bbdcf4f-xflpt\" (UID: \"70231056-c93b-47e7-a2f0-a44329e6f033\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.106734 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.106749 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v2d5\" (UniqueName: \"kubernetes.io/projected/3f64391a-1d93-4e10-9a9a-0b878bde535b-kube-api-access-5v2d5\") pod \"console-operator-58897d9998-pjz2t\" (UID: \"3f64391a-1d93-4e10-9a9a-0b878bde535b\") " pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.106797 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d17e7154-7a80-4607-84d7-016ff6eb07c5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-l2l6r\" (UID: \"d17e7154-7a80-4607-84d7-016ff6eb07c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.106819 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sljh\" (UniqueName: \"kubernetes.io/projected/4b08b51b-daec-4851-99d2-7718232922ea-kube-api-access-5sljh\") pod \"dns-default-rds4c\" (UID: \"4b08b51b-daec-4851-99d2-7718232922ea\") " pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.106849 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-trusted-ca-bundle\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.106893 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.106919 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.106941 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.106960 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-mountpoint-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.106993 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlf55\" (UniqueName: \"kubernetes.io/projected/216b2020-4266-4895-85cb-f0939118b7a8-kube-api-access-mlf55\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107015 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m76cx\" (UniqueName: \"kubernetes.io/projected/14215523-611c-4c7d-a2a9-f4dbe478b0e0-kube-api-access-m76cx\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107033 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/570effcf-7e57-4b00-b75b-a21841431c2b-certs\") pod \"machine-config-server-fcggf\" (UID: \"570effcf-7e57-4b00-b75b-a21841431c2b\") " pod="openshift-machine-config-operator/machine-config-server-fcggf" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107053 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/da22505d-c178-4739-ac34-0d6750469f7b-proxy-tls\") pod \"machine-config-controller-84d6567774-2p8wm\" (UID: \"da22505d-c178-4739-ac34-0d6750469f7b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107075 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2a978e44-b146-45e5-85e0-d86099cbfec0-apiservice-cert\") pod \"packageserver-d55dfcdfc-d9nnn\" (UID: \"2a978e44-b146-45e5-85e0-d86099cbfec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107130 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1b2217f7-f478-4833-b301-097e7c1c07c9-registry-certificates\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107148 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-config\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107165 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-plugins-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107202 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3f64391a-1d93-4e10-9a9a-0b878bde535b-trusted-ca\") pod \"console-operator-58897d9998-pjz2t\" (UID: \"3f64391a-1d93-4e10-9a9a-0b878bde535b\") " pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107224 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zklvg\" (UID: \"7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107246 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107265 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk8mt\" (UniqueName: \"kubernetes.io/projected/bb805699-c562-4a60-a45a-67dbbba49779-kube-api-access-vk8mt\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107285 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/570effcf-7e57-4b00-b75b-a21841431c2b-node-bootstrap-token\") pod \"machine-config-server-fcggf\" (UID: \"570effcf-7e57-4b00-b75b-a21841431c2b\") " pod="openshift-machine-config-operator/machine-config-server-fcggf" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107343 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4bdr\" (UniqueName: \"kubernetes.io/projected/4a4443bb-3033-4719-b515-982c9f3f776a-kube-api-access-l4bdr\") pod \"machine-config-operator-74547568cd-f8825\" (UID: \"4a4443bb-3033-4719-b515-982c9f3f776a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107364 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb805699-c562-4a60-a45a-67dbbba49779-console-serving-cert\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107403 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c0d48b-d1ce-4325-8d7f-6be4f4a968c9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sgjbq\" (UID: \"32c0d48b-d1ce-4325-8d7f-6be4f4a968c9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107422 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4b08b51b-daec-4851-99d2-7718232922ea-metrics-tls\") pod \"dns-default-rds4c\" (UID: \"4b08b51b-daec-4851-99d2-7718232922ea\") " pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107447 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbh9n\" (UniqueName: \"kubernetes.io/projected/32c0d48b-d1ce-4325-8d7f-6be4f4a968c9-kube-api-access-jbh9n\") pod \"openshift-controller-manager-operator-756b6f6bc6-sgjbq\" (UID: \"32c0d48b-d1ce-4325-8d7f-6be4f4a968c9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107481 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzxv4\" (UniqueName: \"kubernetes.io/projected/2a978e44-b146-45e5-85e0-d86099cbfec0-kube-api-access-zzxv4\") pod \"packageserver-d55dfcdfc-d9nnn\" (UID: \"2a978e44-b146-45e5-85e0-d86099cbfec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107503 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f-serving-cert\") pod \"openshift-config-operator-7777fb866f-zklvg\" (UID: \"7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107522 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2a978e44-b146-45e5-85e0-d86099cbfec0-webhook-cert\") pod \"packageserver-d55dfcdfc-d9nnn\" (UID: \"2a978e44-b146-45e5-85e0-d86099cbfec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107543 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41d15067-9885-4940-80f8-fbd63458c326-config\") pod \"service-ca-operator-777779d784-hflmk\" (UID: \"41d15067-9885-4940-80f8-fbd63458c326\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107561 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj4fx\" (UniqueName: \"kubernetes.io/projected/570effcf-7e57-4b00-b75b-a21841431c2b-kube-api-access-bj4fx\") pod \"machine-config-server-fcggf\" (UID: \"570effcf-7e57-4b00-b75b-a21841431c2b\") " pod="openshift-machine-config-operator/machine-config-server-fcggf" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107585 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4a4443bb-3033-4719-b515-982c9f3f776a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-f8825\" (UID: \"4a4443bb-3033-4719-b515-982c9f3f776a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107637 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d5481818-bc94-4c08-9bfe-444c1ed82e1a-metrics-tls\") pod \"dns-operator-744455d44c-dr8c5\" (UID: \"d5481818-bc94-4c08-9bfe-444c1ed82e1a\") " pod="openshift-dns-operator/dns-operator-744455d44c-dr8c5" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.107658 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70231056-c93b-47e7-a2f0-a44329e6f033-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xflpt\" (UID: \"70231056-c93b-47e7-a2f0-a44329e6f033\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.108331 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70231056-c93b-47e7-a2f0-a44329e6f033-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xflpt\" (UID: \"70231056-c93b-47e7-a2f0-a44329e6f033\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.108761 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-csi-data-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.108844 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqw5g\" (UniqueName: \"kubernetes.io/projected/275d1220-b88e-4ff4-a07d-332cb73bb4fa-kube-api-access-cqw5g\") pod \"ingress-canary-tbc7j\" (UID: \"275d1220-b88e-4ff4-a07d-332cb73bb4fa\") " pod="openshift-ingress-canary/ingress-canary-tbc7j" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.108926 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-registry-tls\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.108952 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-audit-policies\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.108975 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41d15067-9885-4940-80f8-fbd63458c326-serving-cert\") pod \"service-ca-operator-777779d784-hflmk\" (UID: \"41d15067-9885-4940-80f8-fbd63458c326\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109043 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1b2217f7-f478-4833-b301-097e7c1c07c9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109183 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14215523-611c-4c7d-a2a9-f4dbe478b0e0-audit-dir\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109210 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c0d48b-d1ce-4325-8d7f-6be4f4a968c9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sgjbq\" (UID: \"32c0d48b-d1ce-4325-8d7f-6be4f4a968c9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109296 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70231056-c93b-47e7-a2f0-a44329e6f033-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xflpt\" (UID: \"70231056-c93b-47e7-a2f0-a44329e6f033\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109344 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d17e7154-7a80-4607-84d7-016ff6eb07c5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-l2l6r\" (UID: \"d17e7154-7a80-4607-84d7-016ff6eb07c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109404 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/216b2020-4266-4895-85cb-f0939118b7a8-serving-cert\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109426 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dc8p\" (UniqueName: \"kubernetes.io/projected/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-kube-api-access-5dc8p\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109495 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b08b51b-daec-4851-99d2-7718232922ea-config-volume\") pod \"dns-default-rds4c\" (UID: \"4b08b51b-daec-4851-99d2-7718232922ea\") " pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109586 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-registration-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109667 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4a4443bb-3033-4719-b515-982c9f3f776a-proxy-tls\") pod \"machine-config-operator-74547568cd-f8825\" (UID: \"4a4443bb-3033-4719-b515-982c9f3f776a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109723 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109751 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109770 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d17e7154-7a80-4607-84d7-016ff6eb07c5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-l2l6r\" (UID: \"d17e7154-7a80-4607-84d7-016ff6eb07c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109842 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn9jq\" (UniqueName: \"kubernetes.io/projected/3dd5759b-8411-48b6-a537-422f23c7d00d-kube-api-access-dn9jq\") pod \"service-ca-9c57cc56f-zc56g\" (UID: \"3dd5759b-8411-48b6-a537-422f23c7d00d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109894 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-service-ca\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.109919 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2a978e44-b146-45e5-85e0-d86099cbfec0-tmpfs\") pod \"packageserver-d55dfcdfc-d9nnn\" (UID: \"2a978e44-b146-45e5-85e0-d86099cbfec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.110017 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/da22505d-c178-4739-ac34-0d6750469f7b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2p8wm\" (UID: \"da22505d-c178-4739-ac34-0d6750469f7b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.110081 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4a4443bb-3033-4719-b515-982c9f3f776a-images\") pod \"machine-config-operator-74547568cd-f8825\" (UID: \"4a4443bb-3033-4719-b515-982c9f3f776a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.110146 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1b2217f7-f478-4833-b301-097e7c1c07c9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.110171 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3dd5759b-8411-48b6-a537-422f23c7d00d-signing-key\") pod \"service-ca-9c57cc56f-zc56g\" (UID: \"3dd5759b-8411-48b6-a537-422f23c7d00d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.110217 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3dd5759b-8411-48b6-a537-422f23c7d00d-signing-cabundle\") pod \"service-ca-9c57cc56f-zc56g\" (UID: \"3dd5759b-8411-48b6-a537-422f23c7d00d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.110483 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.112339 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41d15067-9885-4940-80f8-fbd63458c326-config\") pod \"service-ca-operator-777779d784-hflmk\" (UID: \"41d15067-9885-4940-80f8-fbd63458c326\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.113597 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b2217f7-f478-4833-b301-097e7c1c07c9-trusted-ca\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.114316 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4a4443bb-3033-4719-b515-982c9f3f776a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-f8825\" (UID: \"4a4443bb-3033-4719-b515-982c9f3f776a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:06 crc kubenswrapper[4744]: E1008 09:14:06.115076 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:06.615058702 +0000 UTC m=+141.862704071 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.115898 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-client-ca\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.115938 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.116012 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f64391a-1d93-4e10-9a9a-0b878bde535b-config\") pod \"console-operator-58897d9998-pjz2t\" (UID: \"3f64391a-1d93-4e10-9a9a-0b878bde535b\") " pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.116107 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.116138 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t88tc\" (UniqueName: \"kubernetes.io/projected/05dd98d8-f557-49ea-84ce-2c23b9de7220-kube-api-access-t88tc\") pod \"downloads-7954f5f757-8whxk\" (UID: \"05dd98d8-f557-49ea-84ce-2c23b9de7220\") " pod="openshift-console/downloads-7954f5f757-8whxk" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.116263 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-bound-sa-token\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.116330 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.116360 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2hpx\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-kube-api-access-s2hpx\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.116402 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc2fp\" (UniqueName: \"kubernetes.io/projected/7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f-kube-api-access-wc2fp\") pod \"openshift-config-operator-7777fb866f-zklvg\" (UID: \"7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.118863 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1b2217f7-f478-4833-b301-097e7c1c07c9-registry-certificates\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.119504 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.120121 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-config\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.122720 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.126361 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.126968 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3dd5759b-8411-48b6-a537-422f23c7d00d-signing-cabundle\") pod \"service-ca-9c57cc56f-zc56g\" (UID: \"3dd5759b-8411-48b6-a537-422f23c7d00d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.133056 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-client-ca\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.139732 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.141474 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3f64391a-1d93-4e10-9a9a-0b878bde535b-trusted-ca\") pod \"console-operator-58897d9998-pjz2t\" (UID: \"3f64391a-1d93-4e10-9a9a-0b878bde535b\") " pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.141893 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-audit-policies\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.145580 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f64391a-1d93-4e10-9a9a-0b878bde535b-config\") pod \"console-operator-58897d9998-pjz2t\" (UID: \"3f64391a-1d93-4e10-9a9a-0b878bde535b\") " pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.148468 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.149734 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/d17e7154-7a80-4607-84d7-016ff6eb07c5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-l2l6r\" (UID: \"d17e7154-7a80-4607-84d7-016ff6eb07c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.150059 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.163105 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d5481818-bc94-4c08-9bfe-444c1ed82e1a-metrics-tls\") pod \"dns-operator-744455d44c-dr8c5\" (UID: \"d5481818-bc94-4c08-9bfe-444c1ed82e1a\") " pod="openshift-dns-operator/dns-operator-744455d44c-dr8c5" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.163567 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.163793 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.163819 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4a4443bb-3033-4719-b515-982c9f3f776a-proxy-tls\") pod \"machine-config-operator-74547568cd-f8825\" (UID: \"4a4443bb-3033-4719-b515-982c9f3f776a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.165346 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3dd5759b-8411-48b6-a537-422f23c7d00d-signing-key\") pod \"service-ca-9c57cc56f-zc56g\" (UID: \"3dd5759b-8411-48b6-a537-422f23c7d00d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.166143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4a4443bb-3033-4719-b515-982c9f3f776a-images\") pod \"machine-config-operator-74547568cd-f8825\" (UID: \"4a4443bb-3033-4719-b515-982c9f3f776a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.166683 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f64391a-1d93-4e10-9a9a-0b878bde535b-serving-cert\") pod \"console-operator-58897d9998-pjz2t\" (UID: \"3f64391a-1d93-4e10-9a9a-0b878bde535b\") " pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.166910 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d17e7154-7a80-4607-84d7-016ff6eb07c5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-l2l6r\" (UID: \"d17e7154-7a80-4607-84d7-016ff6eb07c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.166944 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14215523-611c-4c7d-a2a9-f4dbe478b0e0-audit-dir\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.167093 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-registry-tls\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.167264 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.167880 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1b2217f7-f478-4833-b301-097e7c1c07c9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.167988 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zklvg\" (UID: \"7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.181753 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.183316 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7cvp\" (UniqueName: \"kubernetes.io/projected/d5481818-bc94-4c08-9bfe-444c1ed82e1a-kube-api-access-c7cvp\") pod \"dns-operator-744455d44c-dr8c5\" (UID: \"d5481818-bc94-4c08-9bfe-444c1ed82e1a\") " pod="openshift-dns-operator/dns-operator-744455d44c-dr8c5" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.189345 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1b2217f7-f478-4833-b301-097e7c1c07c9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.190845 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f-serving-cert\") pod \"openshift-config-operator-7777fb866f-zklvg\" (UID: \"7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.193145 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70231056-c93b-47e7-a2f0-a44329e6f033-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xflpt\" (UID: \"70231056-c93b-47e7-a2f0-a44329e6f033\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.193548 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/216b2020-4266-4895-85cb-f0939118b7a8-serving-cert\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.196721 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41d15067-9885-4940-80f8-fbd63458c326-serving-cert\") pod \"service-ca-operator-777779d784-hflmk\" (UID: \"41d15067-9885-4940-80f8-fbd63458c326\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.196837 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn44b\" (UniqueName: \"kubernetes.io/projected/41d15067-9885-4940-80f8-fbd63458c326-kube-api-access-wn44b\") pod \"service-ca-operator-777779d784-hflmk\" (UID: \"41d15067-9885-4940-80f8-fbd63458c326\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.200914 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqc5c\" (UniqueName: \"kubernetes.io/projected/70231056-c93b-47e7-a2f0-a44329e6f033-kube-api-access-vqc5c\") pod \"openshift-apiserver-operator-796bbdcf4f-xflpt\" (UID: \"70231056-c93b-47e7-a2f0-a44329e6f033\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.208788 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcn5g\" (UniqueName: \"kubernetes.io/projected/d17e7154-7a80-4607-84d7-016ff6eb07c5-kube-api-access-dcn5g\") pod \"cluster-image-registry-operator-dc59b4c8b-l2l6r\" (UID: \"d17e7154-7a80-4607-84d7-016ff6eb07c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.213796 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.216954 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217142 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c0d48b-d1ce-4325-8d7f-6be4f4a968c9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sgjbq\" (UID: \"32c0d48b-d1ce-4325-8d7f-6be4f4a968c9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217175 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dc8p\" (UniqueName: \"kubernetes.io/projected/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-kube-api-access-5dc8p\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217206 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b08b51b-daec-4851-99d2-7718232922ea-config-volume\") pod \"dns-default-rds4c\" (UID: \"4b08b51b-daec-4851-99d2-7718232922ea\") " pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217226 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-registration-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217252 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-service-ca\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217271 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2a978e44-b146-45e5-85e0-d86099cbfec0-tmpfs\") pod \"packageserver-d55dfcdfc-d9nnn\" (UID: \"2a978e44-b146-45e5-85e0-d86099cbfec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217289 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/da22505d-c178-4739-ac34-0d6750469f7b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2p8wm\" (UID: \"da22505d-c178-4739-ac34-0d6750469f7b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217361 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-oauth-serving-cert\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217417 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb805699-c562-4a60-a45a-67dbbba49779-console-oauth-config\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217431 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-console-config\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217454 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/275d1220-b88e-4ff4-a07d-332cb73bb4fa-cert\") pod \"ingress-canary-tbc7j\" (UID: \"275d1220-b88e-4ff4-a07d-332cb73bb4fa\") " pod="openshift-ingress-canary/ingress-canary-tbc7j" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217474 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvhfr\" (UniqueName: \"kubernetes.io/projected/da22505d-c178-4739-ac34-0d6750469f7b-kube-api-access-dvhfr\") pod \"machine-config-controller-84d6567774-2p8wm\" (UID: \"da22505d-c178-4739-ac34-0d6750469f7b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217494 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-socket-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217525 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sljh\" (UniqueName: \"kubernetes.io/projected/4b08b51b-daec-4851-99d2-7718232922ea-kube-api-access-5sljh\") pod \"dns-default-rds4c\" (UID: \"4b08b51b-daec-4851-99d2-7718232922ea\") " pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217542 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-trusted-ca-bundle\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217567 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-mountpoint-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217591 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/570effcf-7e57-4b00-b75b-a21841431c2b-certs\") pod \"machine-config-server-fcggf\" (UID: \"570effcf-7e57-4b00-b75b-a21841431c2b\") " pod="openshift-machine-config-operator/machine-config-server-fcggf" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217606 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/da22505d-c178-4739-ac34-0d6750469f7b-proxy-tls\") pod \"machine-config-controller-84d6567774-2p8wm\" (UID: \"da22505d-c178-4739-ac34-0d6750469f7b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217623 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2a978e44-b146-45e5-85e0-d86099cbfec0-apiservice-cert\") pod \"packageserver-d55dfcdfc-d9nnn\" (UID: \"2a978e44-b146-45e5-85e0-d86099cbfec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217645 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-plugins-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217662 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk8mt\" (UniqueName: \"kubernetes.io/projected/bb805699-c562-4a60-a45a-67dbbba49779-kube-api-access-vk8mt\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217677 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/570effcf-7e57-4b00-b75b-a21841431c2b-node-bootstrap-token\") pod \"machine-config-server-fcggf\" (UID: \"570effcf-7e57-4b00-b75b-a21841431c2b\") " pod="openshift-machine-config-operator/machine-config-server-fcggf" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217715 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb805699-c562-4a60-a45a-67dbbba49779-console-serving-cert\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217739 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c0d48b-d1ce-4325-8d7f-6be4f4a968c9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sgjbq\" (UID: \"32c0d48b-d1ce-4325-8d7f-6be4f4a968c9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217759 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4b08b51b-daec-4851-99d2-7718232922ea-metrics-tls\") pod \"dns-default-rds4c\" (UID: \"4b08b51b-daec-4851-99d2-7718232922ea\") " pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217785 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbh9n\" (UniqueName: \"kubernetes.io/projected/32c0d48b-d1ce-4325-8d7f-6be4f4a968c9-kube-api-access-jbh9n\") pod \"openshift-controller-manager-operator-756b6f6bc6-sgjbq\" (UID: \"32c0d48b-d1ce-4325-8d7f-6be4f4a968c9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzxv4\" (UniqueName: \"kubernetes.io/projected/2a978e44-b146-45e5-85e0-d86099cbfec0-kube-api-access-zzxv4\") pod \"packageserver-d55dfcdfc-d9nnn\" (UID: \"2a978e44-b146-45e5-85e0-d86099cbfec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217836 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2a978e44-b146-45e5-85e0-d86099cbfec0-webhook-cert\") pod \"packageserver-d55dfcdfc-d9nnn\" (UID: \"2a978e44-b146-45e5-85e0-d86099cbfec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217851 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj4fx\" (UniqueName: \"kubernetes.io/projected/570effcf-7e57-4b00-b75b-a21841431c2b-kube-api-access-bj4fx\") pod \"machine-config-server-fcggf\" (UID: \"570effcf-7e57-4b00-b75b-a21841431c2b\") " pod="openshift-machine-config-operator/machine-config-server-fcggf" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217881 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-csi-data-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.217897 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqw5g\" (UniqueName: \"kubernetes.io/projected/275d1220-b88e-4ff4-a07d-332cb73bb4fa-kube-api-access-cqw5g\") pod \"ingress-canary-tbc7j\" (UID: \"275d1220-b88e-4ff4-a07d-332cb73bb4fa\") " pod="openshift-ingress-canary/ingress-canary-tbc7j" Oct 08 09:14:06 crc kubenswrapper[4744]: E1008 09:14:06.218117 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:06.718102305 +0000 UTC m=+141.965747544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.218166 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-mountpoint-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.219339 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-trusted-ca-bundle\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.220292 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-plugins-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.222455 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-console-config\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.222477 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b08b51b-daec-4851-99d2-7718232922ea-config-volume\") pod \"dns-default-rds4c\" (UID: \"4b08b51b-daec-4851-99d2-7718232922ea\") " pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.223017 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-oauth-serving-cert\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.223130 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-csi-data-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.223169 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-registration-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.223278 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-service-ca\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.223691 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2a978e44-b146-45e5-85e0-d86099cbfec0-tmpfs\") pod \"packageserver-d55dfcdfc-d9nnn\" (UID: \"2a978e44-b146-45e5-85e0-d86099cbfec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.223716 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-socket-dir\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.224888 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/da22505d-c178-4739-ac34-0d6750469f7b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2p8wm\" (UID: \"da22505d-c178-4739-ac34-0d6750469f7b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.225641 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c0d48b-d1ce-4325-8d7f-6be4f4a968c9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sgjbq\" (UID: \"32c0d48b-d1ce-4325-8d7f-6be4f4a968c9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.226938 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v2d5\" (UniqueName: \"kubernetes.io/projected/3f64391a-1d93-4e10-9a9a-0b878bde535b-kube-api-access-5v2d5\") pod \"console-operator-58897d9998-pjz2t\" (UID: \"3f64391a-1d93-4e10-9a9a-0b878bde535b\") " pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.228633 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2a978e44-b146-45e5-85e0-d86099cbfec0-apiservice-cert\") pod \"packageserver-d55dfcdfc-d9nnn\" (UID: \"2a978e44-b146-45e5-85e0-d86099cbfec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.228641 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb805699-c562-4a60-a45a-67dbbba49779-console-oauth-config\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.230064 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/570effcf-7e57-4b00-b75b-a21841431c2b-certs\") pod \"machine-config-server-fcggf\" (UID: \"570effcf-7e57-4b00-b75b-a21841431c2b\") " pod="openshift-machine-config-operator/machine-config-server-fcggf" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.232493 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/da22505d-c178-4739-ac34-0d6750469f7b-proxy-tls\") pod \"machine-config-controller-84d6567774-2p8wm\" (UID: \"da22505d-c178-4739-ac34-0d6750469f7b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.239188 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb805699-c562-4a60-a45a-67dbbba49779-console-serving-cert\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.240404 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-dr8c5" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.241157 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4b08b51b-daec-4851-99d2-7718232922ea-metrics-tls\") pod \"dns-default-rds4c\" (UID: \"4b08b51b-daec-4851-99d2-7718232922ea\") " pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.241490 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2a978e44-b146-45e5-85e0-d86099cbfec0-webhook-cert\") pod \"packageserver-d55dfcdfc-d9nnn\" (UID: \"2a978e44-b146-45e5-85e0-d86099cbfec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.243015 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c0d48b-d1ce-4325-8d7f-6be4f4a968c9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sgjbq\" (UID: \"32c0d48b-d1ce-4325-8d7f-6be4f4a968c9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.243222 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/570effcf-7e57-4b00-b75b-a21841431c2b-node-bootstrap-token\") pod \"machine-config-server-fcggf\" (UID: \"570effcf-7e57-4b00-b75b-a21841431c2b\") " pod="openshift-machine-config-operator/machine-config-server-fcggf" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.243617 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/275d1220-b88e-4ff4-a07d-332cb73bb4fa-cert\") pod \"ingress-canary-tbc7j\" (UID: \"275d1220-b88e-4ff4-a07d-332cb73bb4fa\") " pod="openshift-ingress-canary/ingress-canary-tbc7j" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.244837 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d17e7154-7a80-4607-84d7-016ff6eb07c5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-l2l6r\" (UID: \"d17e7154-7a80-4607-84d7-016ff6eb07c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.246554 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.261255 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4bdr\" (UniqueName: \"kubernetes.io/projected/4a4443bb-3033-4719-b515-982c9f3f776a-kube-api-access-l4bdr\") pod \"machine-config-operator-74547568cd-f8825\" (UID: \"4a4443bb-3033-4719-b515-982c9f3f776a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.271587 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98"] Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.276902 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" event={"ID":"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83","Type":"ContainerStarted","Data":"12361e6571a87d29a27bbea9cdf6b40eaf1630cb5d1a40143231290a331bebff"} Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.287765 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc2fp\" (UniqueName: \"kubernetes.io/projected/7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f-kube-api-access-wc2fp\") pod \"openshift-config-operator-7777fb866f-zklvg\" (UID: \"7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.289567 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" event={"ID":"fc63fe57-540e-4c29-8fc6-155d220f2e9e","Type":"ContainerStarted","Data":"4631eedd96ca13fd842d90ef32f9c53aaae32e1f30e46d2d8b82492ba67bcff7"} Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.293776 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc"] Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.296421 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fl2qz"] Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.307584 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" event={"ID":"fbc99729-cada-43b3-be3e-528a87db765e","Type":"ContainerStarted","Data":"631c994733fca65e42a1a53043047d0aae7801ef80377c9f0c59f639a6174467"} Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.309467 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" event={"ID":"a7221db0-3f54-4e07-95ec-56c4fb95aecc","Type":"ContainerStarted","Data":"c8a30d46737c0817396209e145c0490b715f28324a11718802e39d516a98bf7e"} Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.310431 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" event={"ID":"05957de0-ed69-46e2-aab9-d146cfcbbb90","Type":"ContainerStarted","Data":"31fd179eb4f26d0474c5f1b9fdacb1a08e37925762f595da5f5abb0c8b6b5af2"} Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.318833 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: E1008 09:14:06.319201 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:06.81918589 +0000 UTC m=+142.066831129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.320783 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlf55\" (UniqueName: \"kubernetes.io/projected/216b2020-4266-4895-85cb-f0939118b7a8-kube-api-access-mlf55\") pod \"controller-manager-879f6c89f-8h78h\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.322673 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m76cx\" (UniqueName: \"kubernetes.io/projected/14215523-611c-4c7d-a2a9-f4dbe478b0e0-kube-api-access-m76cx\") pod \"oauth-openshift-558db77b4-x97sx\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.324083 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5j25g" event={"ID":"23ba3d67-97df-4a13-8603-4c0efa967852","Type":"ContainerStarted","Data":"96b2b0b3845e735bf50e16faf74c8037da481af0e394cef13fc7664bbf03b220"} Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.329216 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" event={"ID":"ff9cb19e-ad39-4636-b302-57b5f39b0072","Type":"ContainerStarted","Data":"4c4a65acb3b3126de5a564faad9818df0a8a4622eddd0e6b71cda52923e9075c"} Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.329264 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" event={"ID":"ff9cb19e-ad39-4636-b302-57b5f39b0072","Type":"ContainerStarted","Data":"0de88f35f62efd38ec83f29221423ed7ec071b20a907500a557a2624fe0e69bc"} Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.329564 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.332094 4744 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-hns6n container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.332148 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" podUID="ff9cb19e-ad39-4636-b302-57b5f39b0072" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.367175 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn9jq\" (UniqueName: \"kubernetes.io/projected/3dd5759b-8411-48b6-a537-422f23c7d00d-kube-api-access-dn9jq\") pod \"service-ca-9c57cc56f-zc56g\" (UID: \"3dd5759b-8411-48b6-a537-422f23c7d00d\") " pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.387182 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t88tc\" (UniqueName: \"kubernetes.io/projected/05dd98d8-f557-49ea-84ce-2c23b9de7220-kube-api-access-t88tc\") pod \"downloads-7954f5f757-8whxk\" (UID: \"05dd98d8-f557-49ea-84ce-2c23b9de7220\") " pod="openshift-console/downloads-7954f5f757-8whxk" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.400712 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv"] Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.407964 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-n26p9"] Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.419589 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.419965 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-bound-sa-token\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: E1008 09:14:06.420207 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:06.920184142 +0000 UTC m=+142.167829381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.426617 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2hpx\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-kube-api-access-s2hpx\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.428527 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.446867 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqw5g\" (UniqueName: \"kubernetes.io/projected/275d1220-b88e-4ff4-a07d-332cb73bb4fa-kube-api-access-cqw5g\") pod \"ingress-canary-tbc7j\" (UID: \"275d1220-b88e-4ff4-a07d-332cb73bb4fa\") " pod="openshift-ingress-canary/ingress-canary-tbc7j" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.449812 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.456812 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.464438 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzxv4\" (UniqueName: \"kubernetes.io/projected/2a978e44-b146-45e5-85e0-d86099cbfec0-kube-api-access-zzxv4\") pod \"packageserver-d55dfcdfc-d9nnn\" (UID: \"2a978e44-b146-45e5-85e0-d86099cbfec0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.479546 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.491180 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.500012 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.507751 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8whxk" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.509007 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvhfr\" (UniqueName: \"kubernetes.io/projected/da22505d-c178-4739-ac34-0d6750469f7b-kube-api-access-dvhfr\") pod \"machine-config-controller-84d6567774-2p8wm\" (UID: \"da22505d-c178-4739-ac34-0d6750469f7b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.510340 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbh9n\" (UniqueName: \"kubernetes.io/projected/32c0d48b-d1ce-4325-8d7f-6be4f4a968c9-kube-api-access-jbh9n\") pod \"openshift-controller-manager-operator-756b6f6bc6-sgjbq\" (UID: \"32c0d48b-d1ce-4325-8d7f-6be4f4a968c9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.521240 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: E1008 09:14:06.522624 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:07.022611376 +0000 UTC m=+142.270256615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.522922 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk8mt\" (UniqueName: \"kubernetes.io/projected/bb805699-c562-4a60-a45a-67dbbba49779-kube-api-access-vk8mt\") pod \"console-f9d7485db-5kssv\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.562037 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.616245 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-tbc7j" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.617111 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.618589 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.619635 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj4fx\" (UniqueName: \"kubernetes.io/projected/570effcf-7e57-4b00-b75b-a21841431c2b-kube-api-access-bj4fx\") pod \"machine-config-server-fcggf\" (UID: \"570effcf-7e57-4b00-b75b-a21841431c2b\") " pod="openshift-machine-config-operator/machine-config-server-fcggf" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.640695 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.641567 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.641601 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fcggf" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.641659 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:06 crc kubenswrapper[4744]: E1008 09:14:06.642989 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:07.142968252 +0000 UTC m=+142.390613491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.646521 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: E1008 09:14:06.648070 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:07.148056473 +0000 UTC m=+142.395701722 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.650735 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dc8p\" (UniqueName: \"kubernetes.io/projected/a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1-kube-api-access-5dc8p\") pod \"csi-hostpathplugin-9knh2\" (UID: \"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.658617 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sljh\" (UniqueName: \"kubernetes.io/projected/4b08b51b-daec-4851-99d2-7718232922ea-kube-api-access-5sljh\") pod \"dns-default-rds4c\" (UID: \"4b08b51b-daec-4851-99d2-7718232922ea\") " pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.714533 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd"] Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.751896 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:06 crc kubenswrapper[4744]: E1008 09:14:06.752212 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:07.252195838 +0000 UTC m=+142.499841077 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.853309 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:06 crc kubenswrapper[4744]: E1008 09:14:06.853966 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:07.353951162 +0000 UTC m=+142.601596401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.884276 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9knh2" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.935612 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:06 crc kubenswrapper[4744]: I1008 09:14:06.960636 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:06 crc kubenswrapper[4744]: E1008 09:14:06.961314 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:07.461298403 +0000 UTC m=+142.708943642 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.064245 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:07 crc kubenswrapper[4744]: E1008 09:14:07.064618 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:07.564601283 +0000 UTC m=+142.812246512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.169913 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:07 crc kubenswrapper[4744]: E1008 09:14:07.173991 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:07.673958173 +0000 UTC m=+142.921603412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.189054 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th"] Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.272341 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:07 crc kubenswrapper[4744]: E1008 09:14:07.272724 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:07.772711059 +0000 UTC m=+143.020356288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.273162 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fjhhp"] Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.288221 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-jgg9g"] Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.320940 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96"] Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.324002 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7"] Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.375619 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:07 crc kubenswrapper[4744]: E1008 09:14:07.375831 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:07.875805983 +0000 UTC m=+143.123451222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.385995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:07 crc kubenswrapper[4744]: E1008 09:14:07.386571 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:07.886553192 +0000 UTC m=+143.134198431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.411408 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fcggf" event={"ID":"570effcf-7e57-4b00-b75b-a21841431c2b","Type":"ContainerStarted","Data":"df8a4b4872b6115ead5d5b308e1e4730f6a0675973843ad874a02ab9dae6a4ab"} Oct 08 09:14:07 crc kubenswrapper[4744]: W1008 09:14:07.439563 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10339d22_99de_4f28_98a2_f3549c0f4c13.slice/crio-18df5bcb476d29c0da113a07c782f6224050ae052b3bb81dd698d9fb50dd8480 WatchSource:0}: Error finding container 18df5bcb476d29c0da113a07c782f6224050ae052b3bb81dd698d9fb50dd8480: Status 404 returned error can't find the container with id 18df5bcb476d29c0da113a07c782f6224050ae052b3bb81dd698d9fb50dd8480 Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.441256 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc" event={"ID":"81cf0144-e970-4fe2-8663-15b006dc1b91","Type":"ContainerStarted","Data":"8140eda503962973ff58227812035555c9a2ee44fe4e65eb0a78335485037c52"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.441308 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc" event={"ID":"81cf0144-e970-4fe2-8663-15b006dc1b91","Type":"ContainerStarted","Data":"0657c9793b83fa61c47e0829b1f3ef81a5348f9cac4661517c920392fa9b1e9e"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.471351 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cf7sw"] Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.505467 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:07 crc kubenswrapper[4744]: E1008 09:14:07.505839 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:08.005812344 +0000 UTC m=+143.253457583 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.506782 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:07 crc kubenswrapper[4744]: E1008 09:14:07.507562 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:08.007543986 +0000 UTC m=+143.255189225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.611482 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:07 crc kubenswrapper[4744]: E1008 09:14:07.620295 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:08.120274915 +0000 UTC m=+143.367920154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.717541 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:07 crc kubenswrapper[4744]: E1008 09:14:07.723017 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:08.222973488 +0000 UTC m=+143.470618727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733168 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" event={"ID":"fc63fe57-540e-4c29-8fc6-155d220f2e9e","Type":"ContainerStarted","Data":"f8a813c710719715df6bf0aefa9882c69704219021a37982d181d488c2c023cb"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733229 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8h78h"] Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733244 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vkldx"] Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733264 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" event={"ID":"bacbc855-fe99-4250-9da3-b64e95ac7c67","Type":"ContainerStarted","Data":"ac4f0742774ae00345863f7dce979932ec68a8860401bf8a5a775c35facb8fab"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733277 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" event={"ID":"bacbc855-fe99-4250-9da3-b64e95ac7c67","Type":"ContainerStarted","Data":"20d7db24c3450f27394145ba10b74c532836785be9e6247ccfd90a0dca033f6c"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733286 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" event={"ID":"62b2562b-f25e-4276-bab1-0b366d60c231","Type":"ContainerStarted","Data":"b17bba309be9653283972ccff27da6958c4474febbce8241e053f46c528ffaa0"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733299 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46"] Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733309 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" event={"ID":"62b2562b-f25e-4276-bab1-0b366d60c231","Type":"ContainerStarted","Data":"4a1c76b6b50894e7f565ed794ed04fa968afd0f0e3c1067f5cffbe91d215f912"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733318 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" event={"ID":"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83","Type":"ContainerStarted","Data":"c1cbff3a7ebbca23d672bd6e0b1d7e0b6451259dbdb603efd064264010d206c0"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733326 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-n26p9" event={"ID":"8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a","Type":"ContainerStarted","Data":"74966ccb8d3116aa3618053e72305b6fe24057785192887d16e9331e42c3f1fd"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733336 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" event={"ID":"05957de0-ed69-46e2-aab9-d146cfcbbb90","Type":"ContainerStarted","Data":"22cb78cc01fb6ae13c3e456fccf7ce3807bc11887ce3af38b75a8d51edc2bdab"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733345 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" event={"ID":"dc5708d4-d312-48a6-931a-b09f0c1dd569","Type":"ContainerStarted","Data":"d4765e6333ebc8989bd872a232922067bcd05ccadcb9d4c4d1d1c079727e25d5"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733357 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" event={"ID":"dc5708d4-d312-48a6-931a-b09f0c1dd569","Type":"ContainerStarted","Data":"c373bf913c93719cb332ed7667d407a82dd6416813ce95ad2bb664b90899aa33"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733381 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5j25g" event={"ID":"23ba3d67-97df-4a13-8603-4c0efa967852","Type":"ContainerStarted","Data":"da03def9f4112df1c61bcf7068f061fefc7db0c7e4b338474e06467081834f5d"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.733444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" event={"ID":"1459a017-1dc3-4189-b81f-0f4dfcee04f4","Type":"ContainerStarted","Data":"644a705666567d7ede78cf7ade412c02b3bf2ab67b515821afde8ad59d79f06f"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.735610 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" event={"ID":"a7221db0-3f54-4e07-95ec-56c4fb95aecc","Type":"ContainerStarted","Data":"291f489c0253c97133522d962804e6711604f32c3c404b68aa355f436f0a4b7f"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.735703 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" event={"ID":"a7221db0-3f54-4e07-95ec-56c4fb95aecc","Type":"ContainerStarted","Data":"00897c1d94c713bed88c4a153a26c269325917fa968d770cf657ac9a8cf93aa0"} Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.747787 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.758625 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5nnfc" podStartSLOduration=122.758606163 podStartE2EDuration="2m2.758606163s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:07.756482941 +0000 UTC m=+143.004128180" watchObservedRunningTime="2025-10-08 09:14:07.758606163 +0000 UTC m=+143.006251402" Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.825235 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:07 crc kubenswrapper[4744]: E1008 09:14:07.826332 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:08.32631643 +0000 UTC m=+143.573961669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.843031 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" podStartSLOduration=122.84218454 podStartE2EDuration="2m2.84218454s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:07.840756328 +0000 UTC m=+143.088401567" watchObservedRunningTime="2025-10-08 09:14:07.84218454 +0000 UTC m=+143.089829779" Oct 08 09:14:07 crc kubenswrapper[4744]: W1008 09:14:07.868436 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod382e1d8e_0916_487b_9f6c_2d946da409f1.slice/crio-93ac3e459004cd6a6c2a7ef990882a71ee0bea13068b397dea6e467747f062fb WatchSource:0}: Error finding container 93ac3e459004cd6a6c2a7ef990882a71ee0bea13068b397dea6e467747f062fb: Status 404 returned error can't find the container with id 93ac3e459004cd6a6c2a7ef990882a71ee0bea13068b397dea6e467747f062fb Oct 08 09:14:07 crc kubenswrapper[4744]: I1008 09:14:07.928091 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:07 crc kubenswrapper[4744]: E1008 09:14:07.938184 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:08.438164594 +0000 UTC m=+143.685809823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.001526 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" podStartSLOduration=123.00150726 podStartE2EDuration="2m3.00150726s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:07.962994839 +0000 UTC m=+143.210640088" watchObservedRunningTime="2025-10-08 09:14:08.00150726 +0000 UTC m=+143.249152509" Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.034176 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:08 crc kubenswrapper[4744]: E1008 09:14:08.034594 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:08.53457868 +0000 UTC m=+143.782223909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.084676 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-m4wbt" podStartSLOduration=123.084654453 podStartE2EDuration="2m3.084654453s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:08.045930956 +0000 UTC m=+143.293576215" watchObservedRunningTime="2025-10-08 09:14:08.084654453 +0000 UTC m=+143.332299682" Oct 08 09:14:08 crc kubenswrapper[4744]: W1008 09:14:08.091729 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d7eea67_1ce4_4963_9a2d_577d18ce34e5.slice/crio-1f724c68f8d59325cfd62238bbee093e2fb0e2479d8660b213234d2ccdfb4ab5 WatchSource:0}: Error finding container 1f724c68f8d59325cfd62238bbee093e2fb0e2479d8660b213234d2ccdfb4ab5: Status 404 returned error can't find the container with id 1f724c68f8d59325cfd62238bbee093e2fb0e2479d8660b213234d2ccdfb4ab5 Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.101281 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.137874 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dr8c5"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.138471 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:08 crc kubenswrapper[4744]: E1008 09:14:08.138908 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:08.63888207 +0000 UTC m=+143.886527309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.240807 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-54nth" podStartSLOduration=123.240784869 podStartE2EDuration="2m3.240784869s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:08.237445929 +0000 UTC m=+143.485091188" watchObservedRunningTime="2025-10-08 09:14:08.240784869 +0000 UTC m=+143.488430108" Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.241565 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:08 crc kubenswrapper[4744]: E1008 09:14:08.247508 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:08.747487457 +0000 UTC m=+143.995132686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.270666 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:08 crc kubenswrapper[4744]: E1008 09:14:08.276747 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:08.776692262 +0000 UTC m=+144.024337501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.361513 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-b5q8z" podStartSLOduration=123.361493915 podStartE2EDuration="2m3.361493915s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:08.361087142 +0000 UTC m=+143.608732391" watchObservedRunningTime="2025-10-08 09:14:08.361493915 +0000 UTC m=+143.609139144" Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.373254 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:08 crc kubenswrapper[4744]: E1008 09:14:08.374891 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:08.8748441 +0000 UTC m=+144.122489339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.384109 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hflmk"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.452024 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-fl2qz" podStartSLOduration=123.451989866 podStartE2EDuration="2m3.451989866s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:08.429690235 +0000 UTC m=+143.677335484" watchObservedRunningTime="2025-10-08 09:14:08.451989866 +0000 UTC m=+143.699635105" Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.477629 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:08 crc kubenswrapper[4744]: E1008 09:14:08.478123 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:08.978103929 +0000 UTC m=+144.225749168 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.583238 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:08 crc kubenswrapper[4744]: E1008 09:14:08.583680 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:09.083664437 +0000 UTC m=+144.331309676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.638223 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8whxk"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.664618 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.679557 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:08 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:08 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:08 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.679785 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.691741 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:08 crc kubenswrapper[4744]: E1008 09:14:08.692208 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:09.192192902 +0000 UTC m=+144.439838141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.702840 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pjz2t"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.722905 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zc56g"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.747498 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.751316 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x97sx"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.752384 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5kssv"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.752602 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-5j25g" podStartSLOduration=123.752576011 podStartE2EDuration="2m3.752576011s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:08.696266182 +0000 UTC m=+143.943911431" watchObservedRunningTime="2025-10-08 09:14:08.752576011 +0000 UTC m=+144.000221250" Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.760646 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r"] Oct 08 09:14:08 crc kubenswrapper[4744]: W1008 09:14:08.765495 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb805699_c562_4a60_a45a_67dbbba49779.slice/crio-4656b1c3d2f280cb69f6fd396de0bdddb2cf40d182e65b95c2d905ea803cb303 WatchSource:0}: Error finding container 4656b1c3d2f280cb69f6fd396de0bdddb2cf40d182e65b95c2d905ea803cb303: Status 404 returned error can't find the container with id 4656b1c3d2f280cb69f6fd396de0bdddb2cf40d182e65b95c2d905ea803cb303 Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.794438 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:08 crc kubenswrapper[4744]: E1008 09:14:08.794841 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:09.294823432 +0000 UTC m=+144.542468671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.795413 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9knh2"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.810779 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" event={"ID":"18e8bf6d-bea3-435c-9e4b-4d7a287f0c83","Type":"ContainerStarted","Data":"fe695c5e8670f3ce1490c780aca625d356dc176de4c49975412104b09b2c48ee"} Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.911784 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-f8825"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.911837 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" event={"ID":"382e1d8e-0916-487b-9f6c-2d946da409f1","Type":"ContainerStarted","Data":"397ad7ee1c10dd3e48ac2b8fa0931662bde2f88a10f857d0865d6838def22ad6"} Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.911860 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" event={"ID":"382e1d8e-0916-487b-9f6c-2d946da409f1","Type":"ContainerStarted","Data":"93ac3e459004cd6a6c2a7ef990882a71ee0bea13068b397dea6e467747f062fb"} Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.914554 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.916751 4744 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vkldx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.916826 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" podUID="382e1d8e-0916-487b-9f6c-2d946da409f1" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.921857 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:08 crc kubenswrapper[4744]: E1008 09:14:08.924224 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:09.424207856 +0000 UTC m=+144.671853095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.941354 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zklvg"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.965801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" event={"ID":"1459a017-1dc3-4189-b81f-0f4dfcee04f4","Type":"ContainerStarted","Data":"bf29031cbc7c308643b1bc61398e8d94dd7acb1606d5e8b330b5cb1713d3afdf"} Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.966506 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.975469 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq"] Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.980054 4744 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-tfgxd container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.980160 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" podUID="1459a017-1dc3-4189-b81f-0f4dfcee04f4" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 08 09:14:08 crc kubenswrapper[4744]: I1008 09:14:08.991291 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" event={"ID":"10339d22-99de-4f28-98a2-f3549c0f4c13","Type":"ContainerStarted","Data":"18df5bcb476d29c0da113a07c782f6224050ae052b3bb81dd698d9fb50dd8480"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.020331 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" event={"ID":"62b2562b-f25e-4276-bab1-0b366d60c231","Type":"ContainerStarted","Data":"f8c0e93c6afe6b7fae8d3c655348f71b1d925025aba93acb65f49bd6bad5fe8a"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.023955 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:09 crc kubenswrapper[4744]: E1008 09:14:09.025021 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:09.524995631 +0000 UTC m=+144.772640860 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.027165 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm"] Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.071347 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" event={"ID":"216b2020-4266-4895-85cb-f0939118b7a8","Type":"ContainerStarted","Data":"1b48011b9277f1094d7e1f74c108f758e6656dec57b2fdc20c175d4df222475c"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.074883 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" event={"ID":"800b4855-f900-4968-9a86-fad43cec8fa6","Type":"ContainerStarted","Data":"f917294b81adec4e5747d4e80bbc4ac21737204c908e6a5d8a2dae8c556722e5"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.090943 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.100530 4744 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-8h78h container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.100587 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" podUID="216b2020-4266-4895-85cb-f0939118b7a8" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.117020 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn"] Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.118140 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-tbc7j"] Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.126894 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:09 crc kubenswrapper[4744]: E1008 09:14:09.127343 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:09.627324593 +0000 UTC m=+144.874969832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.142398 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" event={"ID":"9d7eea67-1ce4-4963-9a2d-577d18ce34e5","Type":"ContainerStarted","Data":"41794bab5ca9e90be7a6967e3af474b642adbe86398cf04678aa244e9e041ab1"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.142748 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" event={"ID":"9d7eea67-1ce4-4963-9a2d-577d18ce34e5","Type":"ContainerStarted","Data":"1f724c68f8d59325cfd62238bbee093e2fb0e2479d8660b213234d2ccdfb4ab5"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.143911 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.148261 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rds4c"] Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.174089 4744 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hxs46 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.174164 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" podUID="9d7eea67-1ce4-4963-9a2d-577d18ce34e5" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.213609 4744 generic.go:334] "Generic (PLEG): container finished" podID="fbc99729-cada-43b3-be3e-528a87db765e" containerID="c2f85118b0556d4d29f07cf276d64aaa34265f0ac5439da24b6a60f04b74420a" exitCode=0 Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.213748 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" event={"ID":"fbc99729-cada-43b3-be3e-528a87db765e","Type":"ContainerDied","Data":"c2f85118b0556d4d29f07cf276d64aaa34265f0ac5439da24b6a60f04b74420a"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.227596 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" podStartSLOduration=124.227569273 podStartE2EDuration="2m4.227569273s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.214870706 +0000 UTC m=+144.462515955" watchObservedRunningTime="2025-10-08 09:14:09.227569273 +0000 UTC m=+144.475214512" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.230280 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:09 crc kubenswrapper[4744]: E1008 09:14:09.230678 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:09.730651254 +0000 UTC m=+144.978296483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.235604 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:09 crc kubenswrapper[4744]: E1008 09:14:09.236022 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:09.736008473 +0000 UTC m=+144.983653712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:09 crc kubenswrapper[4744]: W1008 09:14:09.252672 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod275d1220_b88e_4ff4_a07d_332cb73bb4fa.slice/crio-db8b0ad816b2272067eb49a132465913c0c3a506977bb8c8804c9507a1fb09e2 WatchSource:0}: Error finding container db8b0ad816b2272067eb49a132465913c0c3a506977bb8c8804c9507a1fb09e2: Status 404 returned error can't find the container with id db8b0ad816b2272067eb49a132465913c0c3a506977bb8c8804c9507a1fb09e2 Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.255043 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n9d98" podStartSLOduration=124.255024157 podStartE2EDuration="2m4.255024157s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.253035328 +0000 UTC m=+144.500680587" watchObservedRunningTime="2025-10-08 09:14:09.255024157 +0000 UTC m=+144.502669396" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.298482 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" event={"ID":"41d15067-9885-4940-80f8-fbd63458c326","Type":"ContainerStarted","Data":"84db4ab6795021bcec4e88fa8af2b3bd55bad9c6965a0686d06bf3ec29f2ff65"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.300434 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" podStartSLOduration=124.300422501 podStartE2EDuration="2m4.300422501s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.300213775 +0000 UTC m=+144.547859024" watchObservedRunningTime="2025-10-08 09:14:09.300422501 +0000 UTC m=+144.548067740" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.333593 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" podStartSLOduration=124.333575254 podStartE2EDuration="2m4.333575254s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.331516593 +0000 UTC m=+144.579161842" watchObservedRunningTime="2025-10-08 09:14:09.333575254 +0000 UTC m=+144.581220493" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.339583 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:09 crc kubenswrapper[4744]: E1008 09:14:09.339956 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:09.839940072 +0000 UTC m=+145.087585311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.343534 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jgg9g" event={"ID":"431d0dc1-8b42-47fb-9a90-71e957e3bddc","Type":"ContainerStarted","Data":"0a63bfc999bae9d9b794fb3f8e7599c07d45d4a55a332af780bb2c8e006c7e09"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.343589 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jgg9g" event={"ID":"431d0dc1-8b42-47fb-9a90-71e957e3bddc","Type":"ContainerStarted","Data":"982b29f652409ec28477a2c346257309364a4d346cb3785a402e7f911bbf7885"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.349488 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nn48b" podStartSLOduration=124.349467434 podStartE2EDuration="2m4.349467434s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.34493277 +0000 UTC m=+144.592578019" watchObservedRunningTime="2025-10-08 09:14:09.349467434 +0000 UTC m=+144.597112673" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.367287 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" podStartSLOduration=124.367269562 podStartE2EDuration="2m4.367269562s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.366077076 +0000 UTC m=+144.613722345" watchObservedRunningTime="2025-10-08 09:14:09.367269562 +0000 UTC m=+144.614914801" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.373850 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" event={"ID":"70231056-c93b-47e7-a2f0-a44329e6f033","Type":"ContainerStarted","Data":"a813a4ebb11b73ffa4813d3fd0d8bfa8d7ce7697702dc68423889768ff86e7c3"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.381465 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96" event={"ID":"09b85eb4-ccad-4223-96a4-b53745e6f585","Type":"ContainerStarted","Data":"ee1aa6d9791c0abd433e6853ba8256e2137623263428f6b89b0e11bfe1a24ab0"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.394974 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dr8c5" event={"ID":"d5481818-bc94-4c08-9bfe-444c1ed82e1a","Type":"ContainerStarted","Data":"5728eb0871e4be9b853a7c1c8099cbb8367a3ae59d4bfa951658db1e8cf6f37c"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.415745 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" event={"ID":"7d11557e-c042-4dc0-bb61-3dfc27b309a1","Type":"ContainerStarted","Data":"b2bf59e9120fd8145caaebc7c268c84ded4c76aab58c45abcac1f437897b7415"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.441332 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:09 crc kubenswrapper[4744]: E1008 09:14:09.444266 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:09.944250482 +0000 UTC m=+145.191895721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.488220 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" podStartSLOduration=124.488198735 podStartE2EDuration="2m4.488198735s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.426934899 +0000 UTC m=+144.674580168" watchObservedRunningTime="2025-10-08 09:14:09.488198735 +0000 UTC m=+144.735843974" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.515281 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fcggf" event={"ID":"570effcf-7e57-4b00-b75b-a21841431c2b","Type":"ContainerStarted","Data":"3177d04b7eb519c5958e3afaf00755e189a5758ca1ce9a345ad1ec608524e4c3"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.547080 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:09 crc kubenswrapper[4744]: E1008 09:14:09.548053 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:10.048005696 +0000 UTC m=+145.295650935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.551138 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-n26p9" event={"ID":"8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a","Type":"ContainerStarted","Data":"0d1f83f1c80083538e3f32efbf5e39cf6c613a6d3d2f98a0ff1831408a59acaa"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.555167 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" podStartSLOduration=124.555146468 podStartE2EDuration="2m4.555146468s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.493595984 +0000 UTC m=+144.741241233" watchObservedRunningTime="2025-10-08 09:14:09.555146468 +0000 UTC m=+144.802791707" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.555356 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" podStartSLOduration=124.555352134 podStartE2EDuration="2m4.555352134s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.553345914 +0000 UTC m=+144.800991153" watchObservedRunningTime="2025-10-08 09:14:09.555352134 +0000 UTC m=+144.802997373" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.620398 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jgg9g" podStartSLOduration=124.620338009 podStartE2EDuration="2m4.620338009s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.602333376 +0000 UTC m=+144.849978615" watchObservedRunningTime="2025-10-08 09:14:09.620338009 +0000 UTC m=+144.867983248" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.627726 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-n26p9" podStartSLOduration=124.627711038 podStartE2EDuration="2m4.627711038s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.626158401 +0000 UTC m=+144.873803650" watchObservedRunningTime="2025-10-08 09:14:09.627711038 +0000 UTC m=+144.875356277" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.649876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:09 crc kubenswrapper[4744]: E1008 09:14:09.652234 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:10.152214413 +0000 UTC m=+145.399859652 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.687548 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:09 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:09 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:09 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.687902 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.701904 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" event={"ID":"6d0058d8-63f2-4864-aa42-a9afbdfa721a","Type":"ContainerStarted","Data":"16545fad126b1c3c41ec7ebbfc5d6af459ff4ddea40bbc7f2e17901f7e186e41"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.701938 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" event={"ID":"6d0058d8-63f2-4864-aa42-a9afbdfa721a","Type":"ContainerStarted","Data":"c13cee236453a244b8a37c5ca8fd9b18807c0debb1b6ddf63d94a69fc1228804"} Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.701980 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.742816 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" podStartSLOduration=124.742795187 podStartE2EDuration="2m4.742795187s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.738359475 +0000 UTC m=+144.986004724" watchObservedRunningTime="2025-10-08 09:14:09.742795187 +0000 UTC m=+144.990440426" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.742909 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-fcggf" podStartSLOduration=6.742906161 podStartE2EDuration="6.742906161s" podCreationTimestamp="2025-10-08 09:14:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:09.653601745 +0000 UTC m=+144.901246984" watchObservedRunningTime="2025-10-08 09:14:09.742906161 +0000 UTC m=+144.990551400" Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.750992 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:09 crc kubenswrapper[4744]: E1008 09:14:09.753440 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:10.253420432 +0000 UTC m=+145.501065671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.862129 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:09 crc kubenswrapper[4744]: E1008 09:14:09.862584 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:10.362570855 +0000 UTC m=+145.610216094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:09 crc kubenswrapper[4744]: I1008 09:14:09.965092 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:09 crc kubenswrapper[4744]: E1008 09:14:09.965648 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:10.465632509 +0000 UTC m=+145.713277748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.068917 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:10 crc kubenswrapper[4744]: E1008 09:14:10.069305 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:10.56929068 +0000 UTC m=+145.816935919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.170158 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:10 crc kubenswrapper[4744]: E1008 09:14:10.170561 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:10.670544949 +0000 UTC m=+145.918190188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.271572 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:10 crc kubenswrapper[4744]: E1008 09:14:10.272120 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:10.772108449 +0000 UTC m=+146.019753688 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.373528 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:10 crc kubenswrapper[4744]: E1008 09:14:10.373893 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:10.873878073 +0000 UTC m=+146.121523312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.475641 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:10 crc kubenswrapper[4744]: E1008 09:14:10.476336 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:10.976322089 +0000 UTC m=+146.223967328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.576881 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:10 crc kubenswrapper[4744]: E1008 09:14:10.577333 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:11.077317401 +0000 UTC m=+146.324962640 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.666828 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:10 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:10 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:10 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.666882 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.679251 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:10 crc kubenswrapper[4744]: E1008 09:14:10.679929 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:11.1799184 +0000 UTC m=+146.427563639 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.725312 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-n26p9" event={"ID":"8e6c2a8d-08d0-4504-99e8-cc56c18d6e1a","Type":"ContainerStarted","Data":"7c2fe7fe19619422537275667b8bc7256e0a0ffcc95cc9a5cd59ac9a4582773a"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.733040 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" event={"ID":"ae32f38d-23e4-402e-9475-9ea45ba692f8","Type":"ContainerStarted","Data":"0b9d7e54330753661298862d18f45f76f4c95a91ce23762bfc020cf25f6f5cbc"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.733102 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" event={"ID":"ae32f38d-23e4-402e-9475-9ea45ba692f8","Type":"ContainerStarted","Data":"d35ffc785711615304c6e93826a8cb4f004207fb9f2091430cf8534778aec2e6"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.736510 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pjz2t" event={"ID":"3f64391a-1d93-4e10-9a9a-0b878bde535b","Type":"ContainerStarted","Data":"0b7cc0cdcf7d0d3919206811e9514db0a16100af69622cea664db0e73442f071"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.736548 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pjz2t" event={"ID":"3f64391a-1d93-4e10-9a9a-0b878bde535b","Type":"ContainerStarted","Data":"d755dfee398a7cd8ef811f4a4ef19082c4194bf8162c30ed28ae37e306e7c996"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.737456 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.742117 4744 patch_prober.go:28] interesting pod/console-operator-58897d9998-pjz2t container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/readyz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.742170 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-pjz2t" podUID="3f64391a-1d93-4e10-9a9a-0b878bde535b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/readyz\": dial tcp 10.217.0.29:8443: connect: connection refused" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.749219 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" event={"ID":"14215523-611c-4c7d-a2a9-f4dbe478b0e0","Type":"ContainerStarted","Data":"4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.749269 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" event={"ID":"14215523-611c-4c7d-a2a9-f4dbe478b0e0","Type":"ContainerStarted","Data":"bfe62576cac9deb3b2129069127fbb5a1341becc7b4effc7250ac71dc8341e8f"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.750219 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.752111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" event={"ID":"3dd5759b-8411-48b6-a537-422f23c7d00d","Type":"ContainerStarted","Data":"70286df0257b88519020c729ec469d1e276e8b34895fdfd52b71f448f56366fb"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.752141 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" event={"ID":"3dd5759b-8411-48b6-a537-422f23c7d00d","Type":"ContainerStarted","Data":"cbe2e8ec1b76e24cfc03965a2b4d63e26b79068bbe8f8a46fcfda73ae7f3a5b3"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.753550 4744 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-x97sx container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.32:6443/healthz\": dial tcp 10.217.0.32:6443: connect: connection refused" start-of-body= Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.753593 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" podUID="14215523-611c-4c7d-a2a9-f4dbe478b0e0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.32:6443/healthz\": dial tcp 10.217.0.32:6443: connect: connection refused" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.757782 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" event={"ID":"2a978e44-b146-45e5-85e0-d86099cbfec0","Type":"ContainerStarted","Data":"2befa86ff2c08e122c776c7a1156eeac4a3f08ff9a8dfdabe06c5c8a47dbf545"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.757822 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" event={"ID":"2a978e44-b146-45e5-85e0-d86099cbfec0","Type":"ContainerStarted","Data":"e5bc1b9ee0c0fc707a4b992e42bce936ce7d4b6cb10949e142d475f7ff10abde"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.758986 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.762677 4744 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-d9nnn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" start-of-body= Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.762746 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" podUID="2a978e44-b146-45e5-85e0-d86099cbfec0" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.768323 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" event={"ID":"216b2020-4266-4895-85cb-f0939118b7a8","Type":"ContainerStarted","Data":"cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.769405 4744 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-8h78h container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.769475 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" podUID="216b2020-4266-4895-85cb-f0939118b7a8" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.782335 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:10 crc kubenswrapper[4744]: E1008 09:14:10.782844 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:11.282811798 +0000 UTC m=+146.530457037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.783717 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:10 crc kubenswrapper[4744]: E1008 09:14:10.784115 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:11.284100916 +0000 UTC m=+146.531746155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.785204 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xflpt" event={"ID":"70231056-c93b-47e7-a2f0-a44329e6f033","Type":"ContainerStarted","Data":"f7e917c9793c4dd161dff6a8f9945dad49f83590a5b605d38138e11d97439aff"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.792470 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" event={"ID":"4a4443bb-3033-4719-b515-982c9f3f776a","Type":"ContainerStarted","Data":"b64cc1fc0d8ba1264d7fc2ccf6eb8139436632a2ddc1d3f8162db061b2783eb2"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.792529 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" event={"ID":"4a4443bb-3033-4719-b515-982c9f3f776a","Type":"ContainerStarted","Data":"11826bde55ebb23967a8584dbafe0262ce2f4831065bc476d68f40c7f160acbd"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.792546 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" event={"ID":"4a4443bb-3033-4719-b515-982c9f3f776a","Type":"ContainerStarted","Data":"4a43299c43e55ef92cb500b7cf2175d5d7de5684a5da9e211dad890d9c169259"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.795599 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-zc56g" podStartSLOduration=125.795587927 podStartE2EDuration="2m5.795587927s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:10.795075061 +0000 UTC m=+146.042720310" watchObservedRunningTime="2025-10-08 09:14:10.795587927 +0000 UTC m=+146.043233166" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.797756 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tkvdh" podStartSLOduration=125.797747651 podStartE2EDuration="2m5.797747651s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:10.7666611 +0000 UTC m=+146.014306329" watchObservedRunningTime="2025-10-08 09:14:10.797747651 +0000 UTC m=+146.045392890" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.810888 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8whxk" event={"ID":"05dd98d8-f557-49ea-84ce-2c23b9de7220","Type":"ContainerStarted","Data":"4961ec90e249e6cb5c10a7feab57241180be1bc369f6edccbaca9f4587067c3f"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.810938 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8whxk" event={"ID":"05dd98d8-f557-49ea-84ce-2c23b9de7220","Type":"ContainerStarted","Data":"df7536be99e25195f0f1c8e7ca7c3e7fa27920b27349c91c8244eb7df38e9722"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.811642 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8whxk" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.831546 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5kssv" event={"ID":"bb805699-c562-4a60-a45a-67dbbba49779","Type":"ContainerStarted","Data":"c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.831925 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5kssv" event={"ID":"bb805699-c562-4a60-a45a-67dbbba49779","Type":"ContainerStarted","Data":"4656b1c3d2f280cb69f6fd396de0bdddb2cf40d182e65b95c2d905ea803cb303"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.832403 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-8whxk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.832532 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8whxk" podUID="05dd98d8-f557-49ea-84ce-2c23b9de7220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.861668 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hflmk" event={"ID":"41d15067-9885-4940-80f8-fbd63458c326","Type":"ContainerStarted","Data":"f2b331d05c391d6548550c339b9260b650d9b3efd60ca68246e5eb51591755f6"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.871801 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" podStartSLOduration=125.871769884 podStartE2EDuration="2m5.871769884s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:10.870860947 +0000 UTC m=+146.118506196" watchObservedRunningTime="2025-10-08 09:14:10.871769884 +0000 UTC m=+146.119415123" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.889077 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dr8c5" event={"ID":"d5481818-bc94-4c08-9bfe-444c1ed82e1a","Type":"ContainerStarted","Data":"a1cf5e39dbf57f9473394c80e54d0466d39b1fdbce2698417ec7361794e6424f"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.889132 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dr8c5" event={"ID":"d5481818-bc94-4c08-9bfe-444c1ed82e1a","Type":"ContainerStarted","Data":"59612d7554bb9d54931108e71b6519da2d9ebf182a117bcea6d62f5c1f3d767f"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.891331 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:10 crc kubenswrapper[4744]: E1008 09:14:10.920703 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:11.420680993 +0000 UTC m=+146.668326232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.933097 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" podStartSLOduration=125.933077451 podStartE2EDuration="2m5.933077451s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:10.932795052 +0000 UTC m=+146.180440311" watchObservedRunningTime="2025-10-08 09:14:10.933077451 +0000 UTC m=+146.180722690" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.950929 4744 generic.go:334] "Generic (PLEG): container finished" podID="10339d22-99de-4f28-98a2-f3549c0f4c13" containerID="ffda1ef8095d85eeca25061837b458ba07170559d4f92235cd8b7631533ac241" exitCode=0 Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.951052 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" event={"ID":"10339d22-99de-4f28-98a2-f3549c0f4c13","Type":"ContainerDied","Data":"ffda1ef8095d85eeca25061837b458ba07170559d4f92235cd8b7631533ac241"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.967601 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" event={"ID":"32c0d48b-d1ce-4325-8d7f-6be4f4a968c9","Type":"ContainerStarted","Data":"d450a42fc01ac469d5265e5b526f28d07b1d540a3e6771a739fa135eb6b6f987"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.967650 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" event={"ID":"32c0d48b-d1ce-4325-8d7f-6be4f4a968c9","Type":"ContainerStarted","Data":"3897323b55cc8bb99e63dd75b82a8904641bde9378f6ed1de5a2ce072f6c5e41"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.969900 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-pjz2t" podStartSLOduration=125.969879941 podStartE2EDuration="2m5.969879941s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:10.968677025 +0000 UTC m=+146.216322274" watchObservedRunningTime="2025-10-08 09:14:10.969879941 +0000 UTC m=+146.217525180" Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.998826 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" event={"ID":"da22505d-c178-4739-ac34-0d6750469f7b","Type":"ContainerStarted","Data":"64295de96610cf8c01a44f8aa1bc4714a786436bfee94c51cdc13ed4256e61d6"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.998887 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" event={"ID":"da22505d-c178-4739-ac34-0d6750469f7b","Type":"ContainerStarted","Data":"408c74042bd5200aa040e9c4a5b9cf33a694f8b89af8d50be10dee6c9190a658"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.998903 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" event={"ID":"da22505d-c178-4739-ac34-0d6750469f7b","Type":"ContainerStarted","Data":"3b8e5050254c3d87ca3d67f8ef1cafe0ef6563cd87bc91b49343c01a8ccfa464"} Oct 08 09:14:10 crc kubenswrapper[4744]: I1008 09:14:10.999616 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:11 crc kubenswrapper[4744]: E1008 09:14:11.002026 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:11.502006152 +0000 UTC m=+146.749651391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.008183 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-99rp7" event={"ID":"800b4855-f900-4968-9a86-fad43cec8fa6","Type":"ContainerStarted","Data":"337d06a44ed63abcda4bd83349bcd31d83f67685368262c91139101f972fe781"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.008877 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-dr8c5" podStartSLOduration=126.008857416 podStartE2EDuration="2m6.008857416s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:11.007694941 +0000 UTC m=+146.255340200" watchObservedRunningTime="2025-10-08 09:14:11.008857416 +0000 UTC m=+146.256502645" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.013598 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" event={"ID":"d17e7154-7a80-4607-84d7-016ff6eb07c5","Type":"ContainerStarted","Data":"583bc93bdfccc0ede71789afd508a8d09946508c334755050d64934b9cf9a6d0"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.013631 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" event={"ID":"d17e7154-7a80-4607-84d7-016ff6eb07c5","Type":"ContainerStarted","Data":"d79b7d10252efaf3353cd628f02b23bf7ea7017b79a65d1c4fa72b783b645fcb"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.039331 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-5kssv" podStartSLOduration=126.039313578 podStartE2EDuration="2m6.039313578s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:11.038131582 +0000 UTC m=+146.285776841" watchObservedRunningTime="2025-10-08 09:14:11.039313578 +0000 UTC m=+146.286958817" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.059842 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96" event={"ID":"09b85eb4-ccad-4223-96a4-b53745e6f585","Type":"ContainerStarted","Data":"b69dbf3eefdc04f0016a68620dbc0ec7cf22a6da8cbf638ae0ca74267d9a0b62"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.059885 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96" event={"ID":"09b85eb4-ccad-4223-96a4-b53745e6f585","Type":"ContainerStarted","Data":"ed9b444b10d4ba2579318d6e77ad87b01f31792eef75a45f56ee5456155477bb"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.083351 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8825" podStartSLOduration=126.083337112 podStartE2EDuration="2m6.083337112s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:11.081081155 +0000 UTC m=+146.328726404" watchObservedRunningTime="2025-10-08 09:14:11.083337112 +0000 UTC m=+146.330982351" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.086602 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" event={"ID":"6d0058d8-63f2-4864-aa42-a9afbdfa721a","Type":"ContainerStarted","Data":"0bc663ae6d968b6e6d50cffb5ff58a60038ab0eeede0edb915229fd137bdd3af"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.098455 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-tbc7j" event={"ID":"275d1220-b88e-4ff4-a07d-332cb73bb4fa","Type":"ContainerStarted","Data":"66801939ebd59886020fb7a6a7ae22c803ac7aec3175a7775a94eaa567b45242"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.098498 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-tbc7j" event={"ID":"275d1220-b88e-4ff4-a07d-332cb73bb4fa","Type":"ContainerStarted","Data":"db8b0ad816b2272067eb49a132465913c0c3a506977bb8c8804c9507a1fb09e2"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.102083 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:11 crc kubenswrapper[4744]: E1008 09:14:11.103613 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:11.603595001 +0000 UTC m=+146.851240240 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.151757 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9knh2" event={"ID":"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1","Type":"ContainerStarted","Data":"7a484f99d4ffb17a9483308383632557a9ed7728b72bd954cf49ff10463cdad5"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.184418 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-8whxk" podStartSLOduration=126.184402816 podStartE2EDuration="2m6.184402816s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:11.183387506 +0000 UTC m=+146.431032765" watchObservedRunningTime="2025-10-08 09:14:11.184402816 +0000 UTC m=+146.432048045" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.196012 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" event={"ID":"fbc99729-cada-43b3-be3e-528a87db765e","Type":"ContainerStarted","Data":"fcba4c45c864c8cb5d2649ce5506aef45eb5d7424ca8d4310303d705a093a1db"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.204326 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:11 crc kubenswrapper[4744]: E1008 09:14:11.204634 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:11.704622575 +0000 UTC m=+146.952267814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.226672 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" event={"ID":"7d11557e-c042-4dc0-bb61-3dfc27b309a1","Type":"ContainerStarted","Data":"ca9425491f6491cd725209a0a9bfe7a7b2e2e39d7e19433e303ba8cbfa32f9ab"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.245715 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jgg9g" event={"ID":"431d0dc1-8b42-47fb-9a90-71e957e3bddc","Type":"ContainerStarted","Data":"04ae7526d69ac8d517a36341e0f547763d0a13f59ce210af138140cf4f326693"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.261914 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sgjbq" podStartSLOduration=126.261895742 podStartE2EDuration="2m6.261895742s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:11.260556062 +0000 UTC m=+146.508201301" watchObservedRunningTime="2025-10-08 09:14:11.261895742 +0000 UTC m=+146.509540981" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.273004 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rds4c" event={"ID":"4b08b51b-daec-4851-99d2-7718232922ea","Type":"ContainerStarted","Data":"691e3fa67d3558043ea0b94fb1ea94a6e291be9448bf078b9f05447cff3017b2"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.273054 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rds4c" event={"ID":"4b08b51b-daec-4851-99d2-7718232922ea","Type":"ContainerStarted","Data":"a8787779cf7d79db4ace8913641181453e7950d24b1acc8df3c281298b858545"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.273844 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.284863 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f" containerID="04f3e0f0e7c71fe807e0a523184a0a79e7cf069a729c5a64b34b74c904c50e4a" exitCode=0 Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.286356 4744 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vkldx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.286410 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" podUID="382e1d8e-0916-487b-9f6c-2d946da409f1" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.286717 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" event={"ID":"7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f","Type":"ContainerDied","Data":"04f3e0f0e7c71fe807e0a523184a0a79e7cf069a729c5a64b34b74c904c50e4a"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.286793 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" event={"ID":"7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f","Type":"ContainerStarted","Data":"b8891c405fd51eaea56d04ea8219af2b3f8e6858e35552beb857b218e634a1f6"} Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.297344 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tfgxd" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.309127 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:11 crc kubenswrapper[4744]: E1008 09:14:11.310350 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:11.810310676 +0000 UTC m=+147.057955915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.355770 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hxs46" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.414401 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:11 crc kubenswrapper[4744]: E1008 09:14:11.415495 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:11.915476862 +0000 UTC m=+147.163122101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.497909 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b6x96" podStartSLOduration=126.497888143 podStartE2EDuration="2m6.497888143s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:11.468922195 +0000 UTC m=+146.716567454" watchObservedRunningTime="2025-10-08 09:14:11.497888143 +0000 UTC m=+146.745533382" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.499304 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rsqj4"] Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.500548 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.517355 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:11 crc kubenswrapper[4744]: E1008 09:14:11.517706 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:12.017687339 +0000 UTC m=+147.265332578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.546174 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.576045 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rsqj4"] Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.606475 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-tbc7j" podStartSLOduration=8.60645893 podStartE2EDuration="8.60645893s" podCreationTimestamp="2025-10-08 09:14:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:11.60375659 +0000 UTC m=+146.851401849" watchObservedRunningTime="2025-10-08 09:14:11.60645893 +0000 UTC m=+146.854104169" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.619382 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-catalog-content\") pod \"community-operators-rsqj4\" (UID: \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\") " pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.619637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr262\" (UniqueName: \"kubernetes.io/projected/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-kube-api-access-zr262\") pod \"community-operators-rsqj4\" (UID: \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\") " pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.619756 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.619865 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-utilities\") pod \"community-operators-rsqj4\" (UID: \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\") " pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:14:11 crc kubenswrapper[4744]: E1008 09:14:11.620290 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:12.120264979 +0000 UTC m=+147.367910408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.673570 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:11 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:11 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:11 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.673640 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.701785 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l2l6r" podStartSLOduration=126.701766833 podStartE2EDuration="2m6.701766833s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:11.696490077 +0000 UTC m=+146.944135336" watchObservedRunningTime="2025-10-08 09:14:11.701766833 +0000 UTC m=+146.949412072" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.720695 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:11 crc kubenswrapper[4744]: E1008 09:14:11.720888 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:12.220861499 +0000 UTC m=+147.468506738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.721037 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-catalog-content\") pod \"community-operators-rsqj4\" (UID: \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\") " pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.721116 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr262\" (UniqueName: \"kubernetes.io/projected/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-kube-api-access-zr262\") pod \"community-operators-rsqj4\" (UID: \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\") " pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.721141 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.721176 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-utilities\") pod \"community-operators-rsqj4\" (UID: \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\") " pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.722124 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-utilities\") pod \"community-operators-rsqj4\" (UID: \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\") " pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.722335 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-catalog-content\") pod \"community-operators-rsqj4\" (UID: \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\") " pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:14:11 crc kubenswrapper[4744]: E1008 09:14:11.722795 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:12.222771955 +0000 UTC m=+147.470417294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.729193 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pvc9m"] Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.730182 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:14:11 crc kubenswrapper[4744]: W1008 09:14:11.779280 4744 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Oct 08 09:14:11 crc kubenswrapper[4744]: E1008 09:14:11.780051 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.823232 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.823456 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62e5fd77-7efc-455d-86d0-763bb040f2d0-utilities\") pod \"certified-operators-pvc9m\" (UID: \"62e5fd77-7efc-455d-86d0-763bb040f2d0\") " pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.823496 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62e5fd77-7efc-455d-86d0-763bb040f2d0-catalog-content\") pod \"certified-operators-pvc9m\" (UID: \"62e5fd77-7efc-455d-86d0-763bb040f2d0\") " pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.823583 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5jx6\" (UniqueName: \"kubernetes.io/projected/62e5fd77-7efc-455d-86d0-763bb040f2d0-kube-api-access-w5jx6\") pod \"certified-operators-pvc9m\" (UID: \"62e5fd77-7efc-455d-86d0-763bb040f2d0\") " pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:14:11 crc kubenswrapper[4744]: E1008 09:14:11.823741 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:12.323722647 +0000 UTC m=+147.571367886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.874473 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr262\" (UniqueName: \"kubernetes.io/projected/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-kube-api-access-zr262\") pod \"community-operators-rsqj4\" (UID: \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\") " pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.876658 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pvc9m"] Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.925308 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62e5fd77-7efc-455d-86d0-763bb040f2d0-catalog-content\") pod \"certified-operators-pvc9m\" (UID: \"62e5fd77-7efc-455d-86d0-763bb040f2d0\") " pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.925746 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5jx6\" (UniqueName: \"kubernetes.io/projected/62e5fd77-7efc-455d-86d0-763bb040f2d0-kube-api-access-w5jx6\") pod \"certified-operators-pvc9m\" (UID: \"62e5fd77-7efc-455d-86d0-763bb040f2d0\") " pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.925842 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.925951 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62e5fd77-7efc-455d-86d0-763bb040f2d0-utilities\") pod \"certified-operators-pvc9m\" (UID: \"62e5fd77-7efc-455d-86d0-763bb040f2d0\") " pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.926532 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62e5fd77-7efc-455d-86d0-763bb040f2d0-utilities\") pod \"certified-operators-pvc9m\" (UID: \"62e5fd77-7efc-455d-86d0-763bb040f2d0\") " pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.926850 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62e5fd77-7efc-455d-86d0-763bb040f2d0-catalog-content\") pod \"certified-operators-pvc9m\" (UID: \"62e5fd77-7efc-455d-86d0-763bb040f2d0\") " pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:14:11 crc kubenswrapper[4744]: E1008 09:14:11.927534 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:12.427518912 +0000 UTC m=+147.675164151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.951124 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lnwmc"] Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.953242 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:14:11 crc kubenswrapper[4744]: I1008 09:14:11.956667 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-cf7sw" podStartSLOduration=126.956619313 podStartE2EDuration="2m6.956619313s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:11.929029026 +0000 UTC m=+147.176674275" watchObservedRunningTime="2025-10-08 09:14:11.956619313 +0000 UTC m=+147.204264552" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.012712 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5jx6\" (UniqueName: \"kubernetes.io/projected/62e5fd77-7efc-455d-86d0-763bb040f2d0-kube-api-access-w5jx6\") pod \"certified-operators-pvc9m\" (UID: \"62e5fd77-7efc-455d-86d0-763bb040f2d0\") " pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.015408 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2p8wm" podStartSLOduration=127.015383174 podStartE2EDuration="2m7.015383174s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:12.012935522 +0000 UTC m=+147.260580771" watchObservedRunningTime="2025-10-08 09:14:12.015383174 +0000 UTC m=+147.263028403" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.015969 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lnwmc"] Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.027571 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:12 crc kubenswrapper[4744]: E1008 09:14:12.027803 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:12.527759481 +0000 UTC m=+147.775404720 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.027888 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ssmg\" (UniqueName: \"kubernetes.io/projected/e33828e9-5ea7-4f7d-ab35-474268801800-kube-api-access-6ssmg\") pod \"community-operators-lnwmc\" (UID: \"e33828e9-5ea7-4f7d-ab35-474268801800\") " pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.027935 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e33828e9-5ea7-4f7d-ab35-474268801800-utilities\") pod \"community-operators-lnwmc\" (UID: \"e33828e9-5ea7-4f7d-ab35-474268801800\") " pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.027966 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.027996 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e33828e9-5ea7-4f7d-ab35-474268801800-catalog-content\") pod \"community-operators-lnwmc\" (UID: \"e33828e9-5ea7-4f7d-ab35-474268801800\") " pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:14:12 crc kubenswrapper[4744]: E1008 09:14:12.028438 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:12.528423601 +0000 UTC m=+147.776068840 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.038269 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n7rwc"] Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.040337 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.096127 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" podStartSLOduration=127.096103976 podStartE2EDuration="2m7.096103976s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:12.092938442 +0000 UTC m=+147.340583701" watchObservedRunningTime="2025-10-08 09:14:12.096103976 +0000 UTC m=+147.343749215" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.129733 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.129915 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e33828e9-5ea7-4f7d-ab35-474268801800-catalog-content\") pod \"community-operators-lnwmc\" (UID: \"e33828e9-5ea7-4f7d-ab35-474268801800\") " pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.129970 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gl2n\" (UniqueName: \"kubernetes.io/projected/baee305b-d6aa-41ac-a360-77ceb69f5db0-kube-api-access-5gl2n\") pod \"certified-operators-n7rwc\" (UID: \"baee305b-d6aa-41ac-a360-77ceb69f5db0\") " pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.130090 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baee305b-d6aa-41ac-a360-77ceb69f5db0-utilities\") pod \"certified-operators-n7rwc\" (UID: \"baee305b-d6aa-41ac-a360-77ceb69f5db0\") " pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.130159 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ssmg\" (UniqueName: \"kubernetes.io/projected/e33828e9-5ea7-4f7d-ab35-474268801800-kube-api-access-6ssmg\") pod \"community-operators-lnwmc\" (UID: \"e33828e9-5ea7-4f7d-ab35-474268801800\") " pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.130210 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baee305b-d6aa-41ac-a360-77ceb69f5db0-catalog-content\") pod \"certified-operators-n7rwc\" (UID: \"baee305b-d6aa-41ac-a360-77ceb69f5db0\") " pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.130239 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e33828e9-5ea7-4f7d-ab35-474268801800-utilities\") pod \"community-operators-lnwmc\" (UID: \"e33828e9-5ea7-4f7d-ab35-474268801800\") " pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.130747 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e33828e9-5ea7-4f7d-ab35-474268801800-catalog-content\") pod \"community-operators-lnwmc\" (UID: \"e33828e9-5ea7-4f7d-ab35-474268801800\") " pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.130825 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e33828e9-5ea7-4f7d-ab35-474268801800-utilities\") pod \"community-operators-lnwmc\" (UID: \"e33828e9-5ea7-4f7d-ab35-474268801800\") " pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:14:12 crc kubenswrapper[4744]: E1008 09:14:12.130865 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:12.630839495 +0000 UTC m=+147.878484734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.138201 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n7rwc"] Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.138757 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.164102 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ssmg\" (UniqueName: \"kubernetes.io/projected/e33828e9-5ea7-4f7d-ab35-474268801800-kube-api-access-6ssmg\") pod \"community-operators-lnwmc\" (UID: \"e33828e9-5ea7-4f7d-ab35-474268801800\") " pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.239155 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baee305b-d6aa-41ac-a360-77ceb69f5db0-utilities\") pod \"certified-operators-n7rwc\" (UID: \"baee305b-d6aa-41ac-a360-77ceb69f5db0\") " pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.243955 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baee305b-d6aa-41ac-a360-77ceb69f5db0-catalog-content\") pod \"certified-operators-n7rwc\" (UID: \"baee305b-d6aa-41ac-a360-77ceb69f5db0\") " pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.244083 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.244254 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gl2n\" (UniqueName: \"kubernetes.io/projected/baee305b-d6aa-41ac-a360-77ceb69f5db0-kube-api-access-5gl2n\") pod \"certified-operators-n7rwc\" (UID: \"baee305b-d6aa-41ac-a360-77ceb69f5db0\") " pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.240246 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baee305b-d6aa-41ac-a360-77ceb69f5db0-utilities\") pod \"certified-operators-n7rwc\" (UID: \"baee305b-d6aa-41ac-a360-77ceb69f5db0\") " pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.246210 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baee305b-d6aa-41ac-a360-77ceb69f5db0-catalog-content\") pod \"certified-operators-n7rwc\" (UID: \"baee305b-d6aa-41ac-a360-77ceb69f5db0\") " pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:14:12 crc kubenswrapper[4744]: E1008 09:14:12.246450 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:12.746424689 +0000 UTC m=+147.994070118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.266569 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.303574 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rds4c" event={"ID":"4b08b51b-daec-4851-99d2-7718232922ea","Type":"ContainerStarted","Data":"fb8d0efc0146e2c796c78f2d537cc84df89261f2cf635a2e45f13f512bd27534"} Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.318459 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" event={"ID":"7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f","Type":"ContainerStarted","Data":"76537ce2afd0222e3d50d5a2dc3abce34d1595b2f3a2731c6bb5a7861658b3b8"} Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.319403 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.329554 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9knh2" event={"ID":"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1","Type":"ContainerStarted","Data":"ec43530583d22fd0d75ecf4055a78eb8b6f36d3c2b92e1a5b805c9b8e3f801a0"} Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.337940 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" event={"ID":"10339d22-99de-4f28-98a2-f3549c0f4c13","Type":"ContainerStarted","Data":"cc3919d902bb9521eac7e3cd0ce8328e358cf03967fcf17523202a1f5a0a5322"} Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.339160 4744 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-x97sx container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.32:6443/healthz\": dial tcp 10.217.0.32:6443: connect: connection refused" start-of-body= Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.339203 4744 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vkldx container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.339217 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" podUID="14215523-611c-4c7d-a2a9-f4dbe478b0e0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.32:6443/healthz\": dial tcp 10.217.0.32:6443: connect: connection refused" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.339248 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" podUID="382e1d8e-0916-487b-9f6c-2d946da409f1" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.20:8080/healthz\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.345495 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-8whxk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.345550 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8whxk" podUID="05dd98d8-f557-49ea-84ce-2c23b9de7220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.347686 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gl2n\" (UniqueName: \"kubernetes.io/projected/baee305b-d6aa-41ac-a360-77ceb69f5db0-kube-api-access-5gl2n\") pod \"certified-operators-n7rwc\" (UID: \"baee305b-d6aa-41ac-a360-77ceb69f5db0\") " pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.349738 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.350355 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:12 crc kubenswrapper[4744]: E1008 09:14:12.350878 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:12.850855423 +0000 UTC m=+148.098500722 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.453821 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:12 crc kubenswrapper[4744]: E1008 09:14:12.475085 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:12.975066143 +0000 UTC m=+148.222711382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.558165 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:12 crc kubenswrapper[4744]: E1008 09:14:12.564922 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.064891854 +0000 UTC m=+148.312537093 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.666276 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:12 crc kubenswrapper[4744]: E1008 09:14:12.666841 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.166821143 +0000 UTC m=+148.414466382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.671714 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:12 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:12 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:12 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.671960 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.767831 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:12 crc kubenswrapper[4744]: E1008 09:14:12.768096 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.268064754 +0000 UTC m=+148.515709993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.768284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:12 crc kubenswrapper[4744]: E1008 09:14:12.768808 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.268799555 +0000 UTC m=+148.516444794 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.800693 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-rds4c" podStartSLOduration=9.800670869 podStartE2EDuration="9.800670869s" podCreationTimestamp="2025-10-08 09:14:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:12.740656441 +0000 UTC m=+147.988301690" watchObservedRunningTime="2025-10-08 09:14:12.800670869 +0000 UTC m=+148.048316108" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.868858 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:12 crc kubenswrapper[4744]: E1008 09:14:12.869216 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.36919882 +0000 UTC m=+148.616844059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.870731 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" podStartSLOduration=127.870702704 podStartE2EDuration="2m7.870702704s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:12.869960402 +0000 UTC m=+148.117605661" watchObservedRunningTime="2025-10-08 09:14:12.870702704 +0000 UTC m=+148.118347943" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.970356 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:12 crc kubenswrapper[4744]: E1008 09:14:12.970714 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.470694566 +0000 UTC m=+148.718339795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.985139 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.989394 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:14:12 crc kubenswrapper[4744]: I1008 09:14:12.995184 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.071323 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:13 crc kubenswrapper[4744]: E1008 09:14:13.072089 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.57207256 +0000 UTC m=+148.819717799 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.149756 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-pjz2t" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.177784 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:13 crc kubenswrapper[4744]: E1008 09:14:13.178199 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.678179153 +0000 UTC m=+148.925824392 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.280327 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:13 crc kubenswrapper[4744]: E1008 09:14:13.285095 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.78506034 +0000 UTC m=+149.032705579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.343321 4744 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-d9nnn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": context deadline exceeded" start-of-body= Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.343401 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" podUID="2a978e44-b146-45e5-85e0-d86099cbfec0" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": context deadline exceeded" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.383688 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:13 crc kubenswrapper[4744]: E1008 09:14:13.384131 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.884117475 +0000 UTC m=+149.131762714 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.420056 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9knh2" event={"ID":"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1","Type":"ContainerStarted","Data":"6d302eac2354a5f182b06c669196382a5da9dce234869aa44e5e38b58133b5c7"} Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.487741 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" event={"ID":"10339d22-99de-4f28-98a2-f3549c0f4c13","Type":"ContainerStarted","Data":"130eeae38b4834ae98395b4f3d1e9974face251e7ca1ae4d7a6ef4aff95a1c0e"} Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.490050 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.490440 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.490536 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.490573 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.490606 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.496522 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:14:13 crc kubenswrapper[4744]: E1008 09:14:13.496656 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:13.996636008 +0000 UTC m=+149.244281247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.504197 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.512037 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.515978 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-8whxk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.516062 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8whxk" podUID="05dd98d8-f557-49ea-84ce-2c23b9de7220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.548019 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.584860 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.598322 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:13 crc kubenswrapper[4744]: E1008 09:14:13.601606 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:14.101592567 +0000 UTC m=+149.349237806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.674013 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:13 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:13 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:13 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.674981 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.675123 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" podStartSLOduration=128.675090014 podStartE2EDuration="2m8.675090014s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:13.67357141 +0000 UTC m=+148.921216669" watchObservedRunningTime="2025-10-08 09:14:13.675090014 +0000 UTC m=+148.922735253" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.703505 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:13 crc kubenswrapper[4744]: E1008 09:14:13.704334 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:14.204311981 +0000 UTC m=+149.451957220 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.722055 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-85kw7"] Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.741188 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.768660 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.774092 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.787584 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.815171 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.815235 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c692\" (UniqueName: \"kubernetes.io/projected/3c40f62a-9591-4187-aaf5-bbf7319d013c-kube-api-access-5c692\") pod \"redhat-marketplace-85kw7\" (UID: \"3c40f62a-9591-4187-aaf5-bbf7319d013c\") " pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.815290 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c40f62a-9591-4187-aaf5-bbf7319d013c-catalog-content\") pod \"redhat-marketplace-85kw7\" (UID: \"3c40f62a-9591-4187-aaf5-bbf7319d013c\") " pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.815322 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c40f62a-9591-4187-aaf5-bbf7319d013c-utilities\") pod \"redhat-marketplace-85kw7\" (UID: \"3c40f62a-9591-4187-aaf5-bbf7319d013c\") " pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:13 crc kubenswrapper[4744]: E1008 09:14:13.837597 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:14.337573418 +0000 UTC m=+149.585218657 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.863608 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-85kw7"] Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.931555 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.931815 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c40f62a-9591-4187-aaf5-bbf7319d013c-utilities\") pod \"redhat-marketplace-85kw7\" (UID: \"3c40f62a-9591-4187-aaf5-bbf7319d013c\") " pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.931913 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c692\" (UniqueName: \"kubernetes.io/projected/3c40f62a-9591-4187-aaf5-bbf7319d013c-kube-api-access-5c692\") pod \"redhat-marketplace-85kw7\" (UID: \"3c40f62a-9591-4187-aaf5-bbf7319d013c\") " pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.931949 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c40f62a-9591-4187-aaf5-bbf7319d013c-catalog-content\") pod \"redhat-marketplace-85kw7\" (UID: \"3c40f62a-9591-4187-aaf5-bbf7319d013c\") " pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.932642 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c40f62a-9591-4187-aaf5-bbf7319d013c-utilities\") pod \"redhat-marketplace-85kw7\" (UID: \"3c40f62a-9591-4187-aaf5-bbf7319d013c\") " pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:13 crc kubenswrapper[4744]: E1008 09:14:13.932715 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:14.432699166 +0000 UTC m=+149.680344405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:13 crc kubenswrapper[4744]: I1008 09:14:13.933066 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c40f62a-9591-4187-aaf5-bbf7319d013c-catalog-content\") pod \"redhat-marketplace-85kw7\" (UID: \"3c40f62a-9591-4187-aaf5-bbf7319d013c\") " pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.051100 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:14 crc kubenswrapper[4744]: E1008 09:14:14.051503 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:14.551488526 +0000 UTC m=+149.799133765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.094796 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c692\" (UniqueName: \"kubernetes.io/projected/3c40f62a-9591-4187-aaf5-bbf7319d013c-kube-api-access-5c692\") pod \"redhat-marketplace-85kw7\" (UID: \"3c40f62a-9591-4187-aaf5-bbf7319d013c\") " pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.141865 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-svt5h"] Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.143162 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.153887 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:14 crc kubenswrapper[4744]: E1008 09:14:14.154458 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:14.654432136 +0000 UTC m=+149.902077375 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.185027 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d9nnn" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.253701 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lnwmc"] Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.258155 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k2zn\" (UniqueName: \"kubernetes.io/projected/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-kube-api-access-2k2zn\") pod \"redhat-marketplace-svt5h\" (UID: \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\") " pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.258217 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-utilities\") pod \"redhat-marketplace-svt5h\" (UID: \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\") " pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.258244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.258281 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-catalog-content\") pod \"redhat-marketplace-svt5h\" (UID: \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\") " pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:14:14 crc kubenswrapper[4744]: E1008 09:14:14.258648 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:14.758635263 +0000 UTC m=+150.006280502 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.311421 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pvc9m"] Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.318752 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-svt5h"] Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.358961 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.359226 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k2zn\" (UniqueName: \"kubernetes.io/projected/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-kube-api-access-2k2zn\") pod \"redhat-marketplace-svt5h\" (UID: \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\") " pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.359273 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-utilities\") pod \"redhat-marketplace-svt5h\" (UID: \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\") " pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.359306 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-catalog-content\") pod \"redhat-marketplace-svt5h\" (UID: \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\") " pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.359747 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-catalog-content\") pod \"redhat-marketplace-svt5h\" (UID: \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\") " pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.360089 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-utilities\") pod \"redhat-marketplace-svt5h\" (UID: \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\") " pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:14:14 crc kubenswrapper[4744]: E1008 09:14:14.360204 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:14.860180901 +0000 UTC m=+150.107826140 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.386041 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.388504 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rsqj4"] Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.463518 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:14 crc kubenswrapper[4744]: E1008 09:14:14.474345 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:14.974301922 +0000 UTC m=+150.221947161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.482286 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k2zn\" (UniqueName: \"kubernetes.io/projected/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-kube-api-access-2k2zn\") pod \"redhat-marketplace-svt5h\" (UID: \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\") " pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.515841 4744 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-x97sx container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.32:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.515951 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" podUID="14215523-611c-4c7d-a2a9-f4dbe478b0e0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.32:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.547054 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.547785 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.561913 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.562209 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.562950 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnwmc" event={"ID":"e33828e9-5ea7-4f7d-ab35-474268801800","Type":"ContainerStarted","Data":"ddf39f5a5a0cae0279a98053d73f57c95861e93fce85192eb4bea007cf971db7"} Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.564243 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.581657 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsqj4" event={"ID":"d28208cc-1278-45d6-9d2e-a65a6d0c94d8","Type":"ContainerStarted","Data":"4a8c42d8769c173dce537f986a3900fdeba1609735b0f9b7f5793dd37e7ea71b"} Oct 08 09:14:14 crc kubenswrapper[4744]: E1008 09:14:14.583525 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:15.083490497 +0000 UTC m=+150.331135736 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.587362 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvc9m" event={"ID":"62e5fd77-7efc-455d-86d0-763bb040f2d0","Type":"ContainerStarted","Data":"5b6a45032810b901a099856881223e0d6809f74970ef7baa6ff6ae145ec84c10"} Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.635298 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.656189 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9knh2" event={"ID":"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1","Type":"ContainerStarted","Data":"a68e3c58b6e1834eb525b5cbf92c7611c5b5be665c6ae21837cc71536846f02e"} Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.674873 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c81e8bf0-59bf-41ff-a992-7bd3fed6a970-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c81e8bf0-59bf-41ff-a992-7bd3fed6a970\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.674913 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c81e8bf0-59bf-41ff-a992-7bd3fed6a970-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c81e8bf0-59bf-41ff-a992-7bd3fed6a970\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.674939 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:14 crc kubenswrapper[4744]: E1008 09:14:14.675244 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:15.175229694 +0000 UTC m=+150.422874923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.681391 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6qmwr"] Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.691152 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:14 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:14 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:14 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.691787 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.692914 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.713477 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.740690 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6qmwr"] Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.783260 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.785494 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.785681 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c81e8bf0-59bf-41ff-a992-7bd3fed6a970-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c81e8bf0-59bf-41ff-a992-7bd3fed6a970\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.785751 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6jjv\" (UniqueName: \"kubernetes.io/projected/432cd400-49d6-4dc1-a40d-182f5c6f4556-kube-api-access-s6jjv\") pod \"redhat-operators-6qmwr\" (UID: \"432cd400-49d6-4dc1-a40d-182f5c6f4556\") " pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.785819 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/432cd400-49d6-4dc1-a40d-182f5c6f4556-utilities\") pod \"redhat-operators-6qmwr\" (UID: \"432cd400-49d6-4dc1-a40d-182f5c6f4556\") " pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.785919 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/432cd400-49d6-4dc1-a40d-182f5c6f4556-catalog-content\") pod \"redhat-operators-6qmwr\" (UID: \"432cd400-49d6-4dc1-a40d-182f5c6f4556\") " pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.786005 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c81e8bf0-59bf-41ff-a992-7bd3fed6a970-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c81e8bf0-59bf-41ff-a992-7bd3fed6a970\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.786084 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c81e8bf0-59bf-41ff-a992-7bd3fed6a970-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c81e8bf0-59bf-41ff-a992-7bd3fed6a970\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 09:14:14 crc kubenswrapper[4744]: E1008 09:14:14.787003 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:15.286970465 +0000 UTC m=+150.534615874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.845921 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c81e8bf0-59bf-41ff-a992-7bd3fed6a970-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c81e8bf0-59bf-41ff-a992-7bd3fed6a970\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.888813 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.888878 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6jjv\" (UniqueName: \"kubernetes.io/projected/432cd400-49d6-4dc1-a40d-182f5c6f4556-kube-api-access-s6jjv\") pod \"redhat-operators-6qmwr\" (UID: \"432cd400-49d6-4dc1-a40d-182f5c6f4556\") " pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.888912 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/432cd400-49d6-4dc1-a40d-182f5c6f4556-utilities\") pod \"redhat-operators-6qmwr\" (UID: \"432cd400-49d6-4dc1-a40d-182f5c6f4556\") " pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.888961 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/432cd400-49d6-4dc1-a40d-182f5c6f4556-catalog-content\") pod \"redhat-operators-6qmwr\" (UID: \"432cd400-49d6-4dc1-a40d-182f5c6f4556\") " pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.889453 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/432cd400-49d6-4dc1-a40d-182f5c6f4556-catalog-content\") pod \"redhat-operators-6qmwr\" (UID: \"432cd400-49d6-4dc1-a40d-182f5c6f4556\") " pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:14:14 crc kubenswrapper[4744]: E1008 09:14:14.889806 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:15.389789041 +0000 UTC m=+150.637434280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.890691 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/432cd400-49d6-4dc1-a40d-182f5c6f4556-utilities\") pod \"redhat-operators-6qmwr\" (UID: \"432cd400-49d6-4dc1-a40d-182f5c6f4556\") " pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.896126 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 09:14:14 crc kubenswrapper[4744]: I1008 09:14:14.993284 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:14 crc kubenswrapper[4744]: E1008 09:14:14.994142 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:15.494120002 +0000 UTC m=+150.741765241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.012558 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6jjv\" (UniqueName: \"kubernetes.io/projected/432cd400-49d6-4dc1-a40d-182f5c6f4556-kube-api-access-s6jjv\") pod \"redhat-operators-6qmwr\" (UID: \"432cd400-49d6-4dc1-a40d-182f5c6f4556\") " pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.065534 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-msvkb"] Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.067285 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.098516 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.099054 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-msvkb"] Oct 08 09:14:15 crc kubenswrapper[4744]: E1008 09:14:15.099204 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:15.599084052 +0000 UTC m=+150.846729291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.135920 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.203438 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.203712 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs2lc\" (UniqueName: \"kubernetes.io/projected/afc8f9d4-5df0-420c-9cba-aae825a33836-kube-api-access-gs2lc\") pod \"redhat-operators-msvkb\" (UID: \"afc8f9d4-5df0-420c-9cba-aae825a33836\") " pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:15 crc kubenswrapper[4744]: E1008 09:14:15.203899 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:15.703877646 +0000 UTC m=+150.951522885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.203940 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8f9d4-5df0-420c-9cba-aae825a33836-catalog-content\") pod \"redhat-operators-msvkb\" (UID: \"afc8f9d4-5df0-420c-9cba-aae825a33836\") " pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.203964 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8f9d4-5df0-420c-9cba-aae825a33836-utilities\") pod \"redhat-operators-msvkb\" (UID: \"afc8f9d4-5df0-420c-9cba-aae825a33836\") " pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.307103 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8f9d4-5df0-420c-9cba-aae825a33836-catalog-content\") pod \"redhat-operators-msvkb\" (UID: \"afc8f9d4-5df0-420c-9cba-aae825a33836\") " pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.307347 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8f9d4-5df0-420c-9cba-aae825a33836-utilities\") pod \"redhat-operators-msvkb\" (UID: \"afc8f9d4-5df0-420c-9cba-aae825a33836\") " pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.307511 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8f9d4-5df0-420c-9cba-aae825a33836-catalog-content\") pod \"redhat-operators-msvkb\" (UID: \"afc8f9d4-5df0-420c-9cba-aae825a33836\") " pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.307510 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs2lc\" (UniqueName: \"kubernetes.io/projected/afc8f9d4-5df0-420c-9cba-aae825a33836-kube-api-access-gs2lc\") pod \"redhat-operators-msvkb\" (UID: \"afc8f9d4-5df0-420c-9cba-aae825a33836\") " pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.307582 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:15 crc kubenswrapper[4744]: E1008 09:14:15.307890 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:15.807876677 +0000 UTC m=+151.055521906 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.308575 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8f9d4-5df0-420c-9cba-aae825a33836-utilities\") pod \"redhat-operators-msvkb\" (UID: \"afc8f9d4-5df0-420c-9cba-aae825a33836\") " pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.371307 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs2lc\" (UniqueName: \"kubernetes.io/projected/afc8f9d4-5df0-420c-9cba-aae825a33836-kube-api-access-gs2lc\") pod \"redhat-operators-msvkb\" (UID: \"afc8f9d4-5df0-420c-9cba-aae825a33836\") " pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.408913 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:15 crc kubenswrapper[4744]: E1008 09:14:15.409640 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:15.909613252 +0000 UTC m=+151.157258491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.446879 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.483684 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n7rwc"] Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.517053 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:15 crc kubenswrapper[4744]: E1008 09:14:15.523187 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:16.023154035 +0000 UTC m=+151.270799274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.563006 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.564086 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.583190 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:15 crc kubenswrapper[4744]: W1008 09:14:15.614528 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbaee305b_d6aa_41ac_a360_77ceb69f5db0.slice/crio-ee28e3b3c0fc42e24fcc1e26d994f3c28fa89614813ab1fd9f7c458a54c6501b WatchSource:0}: Error finding container ee28e3b3c0fc42e24fcc1e26d994f3c28fa89614813ab1fd9f7c458a54c6501b: Status 404 returned error can't find the container with id ee28e3b3c0fc42e24fcc1e26d994f3c28fa89614813ab1fd9f7c458a54c6501b Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.618276 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:15 crc kubenswrapper[4744]: E1008 09:14:15.619622 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:16.119599983 +0000 UTC m=+151.367245222 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.657541 4744 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-zklvg container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.657591 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" podUID="7ed1ddd1-ed73-4bf9-8dad-e40980d1f69f" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.666335 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.687635 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:15 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:15 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:15 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.688085 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.698649 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a2a49079f4e79998b93b29689810b1df95d7d4bfe3ae0326ba0d9940c41b873f"} Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.720658 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:15 crc kubenswrapper[4744]: E1008 09:14:15.721055 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:16.221041348 +0000 UTC m=+151.468686587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.724516 4744 generic.go:334] "Generic (PLEG): container finished" podID="e33828e9-5ea7-4f7d-ab35-474268801800" containerID="b700c2f160368aa32f40a35c20835c669dd389a63912f2372c2f65a62e169352" exitCode=0 Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.724604 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnwmc" event={"ID":"e33828e9-5ea7-4f7d-ab35-474268801800","Type":"ContainerDied","Data":"b700c2f160368aa32f40a35c20835c669dd389a63912f2372c2f65a62e169352"} Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.732775 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.766698 4744 generic.go:334] "Generic (PLEG): container finished" podID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" containerID="152037bb0738eb0e0dfdf4d4d6199a6e2bc250a722a15ceadaf0e0d81fd712db" exitCode=0 Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.766770 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsqj4" event={"ID":"d28208cc-1278-45d6-9d2e-a65a6d0c94d8","Type":"ContainerDied","Data":"152037bb0738eb0e0dfdf4d4d6199a6e2bc250a722a15ceadaf0e0d81fd712db"} Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.795941 4744 generic.go:334] "Generic (PLEG): container finished" podID="62e5fd77-7efc-455d-86d0-763bb040f2d0" containerID="400cc1edc37956ecdf02470cf0483654bd4f02225ee714ec8e71e330d08de4be" exitCode=0 Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.801946 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvc9m" event={"ID":"62e5fd77-7efc-455d-86d0-763bb040f2d0","Type":"ContainerDied","Data":"400cc1edc37956ecdf02470cf0483654bd4f02225ee714ec8e71e330d08de4be"} Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.828991 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-85kw7"] Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.842066 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7ll8m" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.842135 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.844155 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.854034 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.855528 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:15 crc kubenswrapper[4744]: E1008 09:14:15.861917 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:16.361891491 +0000 UTC m=+151.609536730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.960216 4744 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 08 09:14:15 crc kubenswrapper[4744]: I1008 09:14:15.987439 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:15 crc kubenswrapper[4744]: E1008 09:14:15.989552 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:16.489538203 +0000 UTC m=+151.737183432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.099055 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:16 crc kubenswrapper[4744]: E1008 09:14:16.099352 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:16.599337325 +0000 UTC m=+151.846982564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.207940 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:16 crc kubenswrapper[4744]: E1008 09:14:16.208722 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:16.708708166 +0000 UTC m=+151.956353405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.229530 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zklvg" Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.314125 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:16 crc kubenswrapper[4744]: E1008 09:14:16.315711 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:16.815691625 +0000 UTC m=+152.063336874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.417213 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:16 crc kubenswrapper[4744]: E1008 09:14:16.417849 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:16.91783365 +0000 UTC m=+152.165478879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.471480 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.497296 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-svt5h"] Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.512613 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-8whxk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.512666 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8whxk" podUID="05dd98d8-f557-49ea-84ce-2c23b9de7220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.513001 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-8whxk container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.513029 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8whxk" podUID="05dd98d8-f557-49ea-84ce-2c23b9de7220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.525464 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:16 crc kubenswrapper[4744]: E1008 09:14:16.527666 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 09:14:17.027647875 +0000 UTC m=+152.275293114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.629028 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:16 crc kubenswrapper[4744]: E1008 09:14:16.629549 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 09:14:17.129536493 +0000 UTC m=+152.377181732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5lxjl" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.642634 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.643539 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.654693 4744 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-08T09:14:15.960627636Z","Handler":null,"Name":""} Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.675689 4744 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.675727 4744 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.678092 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:16 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:16 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:16 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.678131 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.710019 4744 patch_prober.go:28] interesting pod/console-f9d7485db-5kssv container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.36:8443/health\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.710070 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5kssv" podUID="bb805699-c562-4a60-a45a-67dbbba49779" containerName="console" probeResult="failure" output="Get \"https://10.217.0.36:8443/health\": dial tcp 10.217.0.36:8443: connect: connection refused" Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.736225 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.884156 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.893806 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-msvkb"] Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.949379 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:16 crc kubenswrapper[4744]: I1008 09:14:16.957190 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9knh2" event={"ID":"a97ca7ca-fe8c-407e-b66c-3b2ab24ff0b1","Type":"ContainerStarted","Data":"7da535389da2153c8b6d33aafb2e40ee197edca766e0c07f087d356294b61424"} Oct 08 09:14:16 crc kubenswrapper[4744]: E1008 09:14:16.990712 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbaee305b_d6aa_41ac_a360_77ceb69f5db0.slice/crio-conmon-9c3ae761e81730a0da4906c9dc39fb77f17c2ca8027ca2a411bd2340fabc141b.scope\": RecentStats: unable to find data in memory cache]" Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:16.994739 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.018272 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.009481 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ec0114ba8a358e4a5b98e822995eade30c3b9338feb27c6cdddc7639ca0d6617"} Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.018437 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5e6dcdef4d6f7864585f860acd4fa43d87b8d5f103d88d054ec32f41cb3fabaa"} Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.056806 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f411e51971ced5923aaf783caa1013cf48958494e5927a7c33eee0361ce5b31b"} Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.056867 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"11d0fb451c0970d2c4aafef737cd9661a12c408232b004177cdbc33fd4df31c8"} Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.057523 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.073934 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svt5h" event={"ID":"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc","Type":"ContainerStarted","Data":"4a1de6582f89bb72196fabb35d97c51199834633aa8244ae1b52eeb382ee6fed"} Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.102288 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6qmwr"] Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.111706 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85kw7" event={"ID":"3c40f62a-9591-4187-aaf5-bbf7319d013c","Type":"ContainerStarted","Data":"28bc326074c71f958725ed207f7f6064137213c79eae417113166dc77f9e7d45"} Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.111750 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85kw7" event={"ID":"3c40f62a-9591-4187-aaf5-bbf7319d013c","Type":"ContainerStarted","Data":"278dc9a7571c5088961e5d832cc20ea9e23f61954744d30f59fc33e5c7b13330"} Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.152577 4744 generic.go:334] "Generic (PLEG): container finished" podID="baee305b-d6aa-41ac-a360-77ceb69f5db0" containerID="9c3ae761e81730a0da4906c9dc39fb77f17c2ca8027ca2a411bd2340fabc141b" exitCode=0 Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.152653 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7rwc" event={"ID":"baee305b-d6aa-41ac-a360-77ceb69f5db0","Type":"ContainerDied","Data":"9c3ae761e81730a0da4906c9dc39fb77f17c2ca8027ca2a411bd2340fabc141b"} Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.152683 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7rwc" event={"ID":"baee305b-d6aa-41ac-a360-77ceb69f5db0","Type":"ContainerStarted","Data":"ee28e3b3c0fc42e24fcc1e26d994f3c28fa89614813ab1fd9f7c458a54c6501b"} Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.197642 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"750515e5e119482ad988b709649e3a3c0e3c5e6ede746e406806e41f15c7ed6b"} Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.222652 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-9knh2" podStartSLOduration=14.222637344 podStartE2EDuration="14.222637344s" podCreationTimestamp="2025-10-08 09:14:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:17.2191209 +0000 UTC m=+152.466766169" watchObservedRunningTime="2025-10-08 09:14:17.222637344 +0000 UTC m=+152.470282583" Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.392562 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.446327 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5lxjl\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.513702 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.568479 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.610496 4744 patch_prober.go:28] interesting pod/apiserver-76f77b778f-fjhhp container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 08 09:14:17 crc kubenswrapper[4744]: [+]log ok Oct 08 09:14:17 crc kubenswrapper[4744]: [+]etcd ok Oct 08 09:14:17 crc kubenswrapper[4744]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 08 09:14:17 crc kubenswrapper[4744]: [+]poststarthook/generic-apiserver-start-informers ok Oct 08 09:14:17 crc kubenswrapper[4744]: [+]poststarthook/max-in-flight-filter ok Oct 08 09:14:17 crc kubenswrapper[4744]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 08 09:14:17 crc kubenswrapper[4744]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 08 09:14:17 crc kubenswrapper[4744]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 08 09:14:17 crc kubenswrapper[4744]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 08 09:14:17 crc kubenswrapper[4744]: [+]poststarthook/project.openshift.io-projectcache ok Oct 08 09:14:17 crc kubenswrapper[4744]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 08 09:14:17 crc kubenswrapper[4744]: [+]poststarthook/openshift.io-startinformers ok Oct 08 09:14:17 crc kubenswrapper[4744]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 08 09:14:17 crc kubenswrapper[4744]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 08 09:14:17 crc kubenswrapper[4744]: livez check failed Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.610584 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" podUID="10339d22-99de-4f28-98a2-f3549c0f4c13" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.673156 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:17 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:17 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:17 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:17 crc kubenswrapper[4744]: I1008 09:14:17.673209 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.017961 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.019756 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.032418 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.032703 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.039744 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.090189 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2796b5f-00d8-491e-b4b0-e5ebd034674b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c2796b5f-00d8-491e-b4b0-e5ebd034674b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.095832 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2796b5f-00d8-491e-b4b0-e5ebd034674b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c2796b5f-00d8-491e-b4b0-e5ebd034674b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.200061 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2796b5f-00d8-491e-b4b0-e5ebd034674b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c2796b5f-00d8-491e-b4b0-e5ebd034674b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.200168 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2796b5f-00d8-491e-b4b0-e5ebd034674b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c2796b5f-00d8-491e-b4b0-e5ebd034674b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.200563 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2796b5f-00d8-491e-b4b0-e5ebd034674b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c2796b5f-00d8-491e-b4b0-e5ebd034674b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.282456 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2796b5f-00d8-491e-b4b0-e5ebd034674b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c2796b5f-00d8-491e-b4b0-e5ebd034674b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.384058 4744 generic.go:334] "Generic (PLEG): container finished" podID="afc8f9d4-5df0-420c-9cba-aae825a33836" containerID="fe0f4abdf5948e21dbb672e73f63fd9f930a78dfa3081b67d5b876e3441eaab0" exitCode=0 Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.384196 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msvkb" event={"ID":"afc8f9d4-5df0-420c-9cba-aae825a33836","Type":"ContainerDied","Data":"fe0f4abdf5948e21dbb672e73f63fd9f930a78dfa3081b67d5b876e3441eaab0"} Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.384234 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msvkb" event={"ID":"afc8f9d4-5df0-420c-9cba-aae825a33836","Type":"ContainerStarted","Data":"6c6090ee653702b139b3ae833f54a6208642f6a2ad437f7b12f4899995c53324"} Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.436647 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.439142 4744 generic.go:334] "Generic (PLEG): container finished" podID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" containerID="652ba64bbd4dc9cbd79d3b648ba3465acd69932cf55c2a67c9831b7b42a3b248" exitCode=0 Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.439209 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svt5h" event={"ID":"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc","Type":"ContainerDied","Data":"652ba64bbd4dc9cbd79d3b648ba3465acd69932cf55c2a67c9831b7b42a3b248"} Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.479047 4744 generic.go:334] "Generic (PLEG): container finished" podID="432cd400-49d6-4dc1-a40d-182f5c6f4556" containerID="35477dd9004849232898b38af24d8de0cf816909cbb232e7233bc617a12c7253" exitCode=0 Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.479177 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qmwr" event={"ID":"432cd400-49d6-4dc1-a40d-182f5c6f4556","Type":"ContainerDied","Data":"35477dd9004849232898b38af24d8de0cf816909cbb232e7233bc617a12c7253"} Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.479207 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qmwr" event={"ID":"432cd400-49d6-4dc1-a40d-182f5c6f4556","Type":"ContainerStarted","Data":"c39d76e55f6712baa8d739727d1cc0f743e3663d71c16730573e287b2fd93a9c"} Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.512413 4744 generic.go:334] "Generic (PLEG): container finished" podID="3c40f62a-9591-4187-aaf5-bbf7319d013c" containerID="28bc326074c71f958725ed207f7f6064137213c79eae417113166dc77f9e7d45" exitCode=0 Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.512524 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85kw7" event={"ID":"3c40f62a-9591-4187-aaf5-bbf7319d013c","Type":"ContainerDied","Data":"28bc326074c71f958725ed207f7f6064137213c79eae417113166dc77f9e7d45"} Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.537644 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c81e8bf0-59bf-41ff-a992-7bd3fed6a970","Type":"ContainerStarted","Data":"85c25b5de601b1df43d1220bfb0eabb9027a9446dcff8c4b17e0f106aa58d02a"} Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.640944 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5lxjl"] Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.703932 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:18 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:18 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:18 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:18 crc kubenswrapper[4744]: I1008 09:14:18.704006 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:19 crc kubenswrapper[4744]: I1008 09:14:19.496264 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 09:14:19 crc kubenswrapper[4744]: I1008 09:14:19.607972 4744 generic.go:334] "Generic (PLEG): container finished" podID="c81e8bf0-59bf-41ff-a992-7bd3fed6a970" containerID="91e884d27b061da98f65805ae2afce5342cf738a459e360f4c157536ab4aa9be" exitCode=0 Oct 08 09:14:19 crc kubenswrapper[4744]: I1008 09:14:19.608414 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c81e8bf0-59bf-41ff-a992-7bd3fed6a970","Type":"ContainerDied","Data":"91e884d27b061da98f65805ae2afce5342cf738a459e360f4c157536ab4aa9be"} Oct 08 09:14:19 crc kubenswrapper[4744]: I1008 09:14:19.674907 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:19 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:19 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:19 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:19 crc kubenswrapper[4744]: I1008 09:14:19.674991 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:19 crc kubenswrapper[4744]: I1008 09:14:19.690000 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:14:19 crc kubenswrapper[4744]: I1008 09:14:19.690067 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:14:19 crc kubenswrapper[4744]: I1008 09:14:19.693994 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" event={"ID":"1b2217f7-f478-4833-b301-097e7c1c07c9","Type":"ContainerStarted","Data":"565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb"} Oct 08 09:14:19 crc kubenswrapper[4744]: I1008 09:14:19.694056 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" event={"ID":"1b2217f7-f478-4833-b301-097e7c1c07c9","Type":"ContainerStarted","Data":"931e961dbdc35ba5c0be812dc9fd984ea1eb77eaae512274d70b032eb2dea1f0"} Oct 08 09:14:19 crc kubenswrapper[4744]: I1008 09:14:19.694403 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:19 crc kubenswrapper[4744]: I1008 09:14:19.726058 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" podStartSLOduration=134.726035449 podStartE2EDuration="2m14.726035449s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:19.725833643 +0000 UTC m=+154.973478892" watchObservedRunningTime="2025-10-08 09:14:19.726035449 +0000 UTC m=+154.973680688" Oct 08 09:14:20 crc kubenswrapper[4744]: I1008 09:14:20.666511 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:20 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:20 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:20 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:20 crc kubenswrapper[4744]: I1008 09:14:20.666888 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:20 crc kubenswrapper[4744]: I1008 09:14:20.750354 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c2796b5f-00d8-491e-b4b0-e5ebd034674b","Type":"ContainerStarted","Data":"3b7110a56c041922a6be06bfced9af742009ab765c43b803a48a06df073fa9b2"} Oct 08 09:14:20 crc kubenswrapper[4744]: I1008 09:14:20.849029 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:20 crc kubenswrapper[4744]: I1008 09:14:20.854900 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-fjhhp" Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.326806 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.425086 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c81e8bf0-59bf-41ff-a992-7bd3fed6a970-kubelet-dir\") pod \"c81e8bf0-59bf-41ff-a992-7bd3fed6a970\" (UID: \"c81e8bf0-59bf-41ff-a992-7bd3fed6a970\") " Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.425220 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c81e8bf0-59bf-41ff-a992-7bd3fed6a970-kube-api-access\") pod \"c81e8bf0-59bf-41ff-a992-7bd3fed6a970\" (UID: \"c81e8bf0-59bf-41ff-a992-7bd3fed6a970\") " Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.426503 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c81e8bf0-59bf-41ff-a992-7bd3fed6a970-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c81e8bf0-59bf-41ff-a992-7bd3fed6a970" (UID: "c81e8bf0-59bf-41ff-a992-7bd3fed6a970"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.449593 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c81e8bf0-59bf-41ff-a992-7bd3fed6a970-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c81e8bf0-59bf-41ff-a992-7bd3fed6a970" (UID: "c81e8bf0-59bf-41ff-a992-7bd3fed6a970"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.534103 4744 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c81e8bf0-59bf-41ff-a992-7bd3fed6a970-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.534137 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c81e8bf0-59bf-41ff-a992-7bd3fed6a970-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.667869 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:21 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:21 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:21 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.667997 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.796909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c81e8bf0-59bf-41ff-a992-7bd3fed6a970","Type":"ContainerDied","Data":"85c25b5de601b1df43d1220bfb0eabb9027a9446dcff8c4b17e0f106aa58d02a"} Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.796983 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85c25b5de601b1df43d1220bfb0eabb9027a9446dcff8c4b17e0f106aa58d02a" Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.796992 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.828925 4744 generic.go:334] "Generic (PLEG): container finished" podID="c2796b5f-00d8-491e-b4b0-e5ebd034674b" containerID="01046831b66916f5a692f09ae75bb8dba9aba382b553becf2d14f2f9d66455da" exitCode=0 Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.829176 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c2796b5f-00d8-491e-b4b0-e5ebd034674b","Type":"ContainerDied","Data":"01046831b66916f5a692f09ae75bb8dba9aba382b553becf2d14f2f9d66455da"} Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.839812 4744 generic.go:334] "Generic (PLEG): container finished" podID="bacbc855-fe99-4250-9da3-b64e95ac7c67" containerID="ac4f0742774ae00345863f7dce979932ec68a8860401bf8a5a775c35facb8fab" exitCode=0 Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.840598 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" event={"ID":"bacbc855-fe99-4250-9da3-b64e95ac7c67","Type":"ContainerDied","Data":"ac4f0742774ae00345863f7dce979932ec68a8860401bf8a5a775c35facb8fab"} Oct 08 09:14:21 crc kubenswrapper[4744]: I1008 09:14:21.944042 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-rds4c" Oct 08 09:14:22 crc kubenswrapper[4744]: I1008 09:14:22.676138 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:22 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:22 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:22 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:22 crc kubenswrapper[4744]: I1008 09:14:22.676500 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.405352 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.406871 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.472082 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bacbc855-fe99-4250-9da3-b64e95ac7c67-config-volume\") pod \"bacbc855-fe99-4250-9da3-b64e95ac7c67\" (UID: \"bacbc855-fe99-4250-9da3-b64e95ac7c67\") " Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.474326 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bacbc855-fe99-4250-9da3-b64e95ac7c67-config-volume" (OuterVolumeSpecName: "config-volume") pod "bacbc855-fe99-4250-9da3-b64e95ac7c67" (UID: "bacbc855-fe99-4250-9da3-b64e95ac7c67"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.474438 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdnxn\" (UniqueName: \"kubernetes.io/projected/bacbc855-fe99-4250-9da3-b64e95ac7c67-kube-api-access-wdnxn\") pod \"bacbc855-fe99-4250-9da3-b64e95ac7c67\" (UID: \"bacbc855-fe99-4250-9da3-b64e95ac7c67\") " Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.474487 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bacbc855-fe99-4250-9da3-b64e95ac7c67-secret-volume\") pod \"bacbc855-fe99-4250-9da3-b64e95ac7c67\" (UID: \"bacbc855-fe99-4250-9da3-b64e95ac7c67\") " Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.474558 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2796b5f-00d8-491e-b4b0-e5ebd034674b-kube-api-access\") pod \"c2796b5f-00d8-491e-b4b0-e5ebd034674b\" (UID: \"c2796b5f-00d8-491e-b4b0-e5ebd034674b\") " Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.474593 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2796b5f-00d8-491e-b4b0-e5ebd034674b-kubelet-dir\") pod \"c2796b5f-00d8-491e-b4b0-e5ebd034674b\" (UID: \"c2796b5f-00d8-491e-b4b0-e5ebd034674b\") " Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.474890 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bacbc855-fe99-4250-9da3-b64e95ac7c67-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.474917 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2796b5f-00d8-491e-b4b0-e5ebd034674b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c2796b5f-00d8-491e-b4b0-e5ebd034674b" (UID: "c2796b5f-00d8-491e-b4b0-e5ebd034674b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.483274 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2796b5f-00d8-491e-b4b0-e5ebd034674b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c2796b5f-00d8-491e-b4b0-e5ebd034674b" (UID: "c2796b5f-00d8-491e-b4b0-e5ebd034674b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.483399 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bacbc855-fe99-4250-9da3-b64e95ac7c67-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bacbc855-fe99-4250-9da3-b64e95ac7c67" (UID: "bacbc855-fe99-4250-9da3-b64e95ac7c67"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.485389 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bacbc855-fe99-4250-9da3-b64e95ac7c67-kube-api-access-wdnxn" (OuterVolumeSpecName: "kube-api-access-wdnxn") pod "bacbc855-fe99-4250-9da3-b64e95ac7c67" (UID: "bacbc855-fe99-4250-9da3-b64e95ac7c67"). InnerVolumeSpecName "kube-api-access-wdnxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.577101 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdnxn\" (UniqueName: \"kubernetes.io/projected/bacbc855-fe99-4250-9da3-b64e95ac7c67-kube-api-access-wdnxn\") on node \"crc\" DevicePath \"\"" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.577838 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bacbc855-fe99-4250-9da3-b64e95ac7c67-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.577873 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2796b5f-00d8-491e-b4b0-e5ebd034674b-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.577882 4744 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c2796b5f-00d8-491e-b4b0-e5ebd034674b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.672101 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:23 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:23 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:23 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.672175 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.941746 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c2796b5f-00d8-491e-b4b0-e5ebd034674b","Type":"ContainerDied","Data":"3b7110a56c041922a6be06bfced9af742009ab765c43b803a48a06df073fa9b2"} Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.942154 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b7110a56c041922a6be06bfced9af742009ab765c43b803a48a06df073fa9b2" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.942004 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.952125 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" event={"ID":"bacbc855-fe99-4250-9da3-b64e95ac7c67","Type":"ContainerDied","Data":"20d7db24c3450f27394145ba10b74c532836785be9e6247ccfd90a0dca033f6c"} Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.952164 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20d7db24c3450f27394145ba10b74c532836785be9e6247ccfd90a0dca033f6c" Oct 08 09:14:23 crc kubenswrapper[4744]: I1008 09:14:23.952215 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv" Oct 08 09:14:24 crc kubenswrapper[4744]: I1008 09:14:24.665020 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:24 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:24 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:24 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:24 crc kubenswrapper[4744]: I1008 09:14:24.665111 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:25 crc kubenswrapper[4744]: I1008 09:14:25.666460 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:25 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:25 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:25 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:25 crc kubenswrapper[4744]: I1008 09:14:25.666516 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:26 crc kubenswrapper[4744]: I1008 09:14:26.511395 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-8whxk container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Oct 08 09:14:26 crc kubenswrapper[4744]: I1008 09:14:26.511451 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8whxk" podUID="05dd98d8-f557-49ea-84ce-2c23b9de7220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Oct 08 09:14:26 crc kubenswrapper[4744]: I1008 09:14:26.511561 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-8whxk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Oct 08 09:14:26 crc kubenswrapper[4744]: I1008 09:14:26.511625 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8whxk" podUID="05dd98d8-f557-49ea-84ce-2c23b9de7220" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Oct 08 09:14:26 crc kubenswrapper[4744]: I1008 09:14:26.644774 4744 patch_prober.go:28] interesting pod/console-f9d7485db-5kssv container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.36:8443/health\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Oct 08 09:14:26 crc kubenswrapper[4744]: I1008 09:14:26.644833 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5kssv" podUID="bb805699-c562-4a60-a45a-67dbbba49779" containerName="console" probeResult="failure" output="Get \"https://10.217.0.36:8443/health\": dial tcp 10.217.0.36:8443: connect: connection refused" Oct 08 09:14:26 crc kubenswrapper[4744]: I1008 09:14:26.665662 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:26 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:26 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:26 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:26 crc kubenswrapper[4744]: I1008 09:14:26.665765 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:27 crc kubenswrapper[4744]: I1008 09:14:27.665878 4744 patch_prober.go:28] interesting pod/router-default-5444994796-5j25g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 09:14:27 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 08 09:14:27 crc kubenswrapper[4744]: [+]process-running ok Oct 08 09:14:27 crc kubenswrapper[4744]: healthz check failed Oct 08 09:14:27 crc kubenswrapper[4744]: I1008 09:14:27.666139 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5j25g" podUID="23ba3d67-97df-4a13-8603-4c0efa967852" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 09:14:27 crc kubenswrapper[4744]: I1008 09:14:27.852504 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:14:27 crc kubenswrapper[4744]: I1008 09:14:27.858276 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be364129-0f0b-41d4-b5f8-1b1def9a1c38-metrics-certs\") pod \"network-metrics-daemon-cdcbz\" (UID: \"be364129-0f0b-41d4-b5f8-1b1def9a1c38\") " pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:14:27 crc kubenswrapper[4744]: I1008 09:14:27.973869 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cdcbz" Oct 08 09:14:28 crc kubenswrapper[4744]: I1008 09:14:28.665698 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:28 crc kubenswrapper[4744]: I1008 09:14:28.668312 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-5j25g" Oct 08 09:14:36 crc kubenswrapper[4744]: I1008 09:14:36.516884 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-8whxk" Oct 08 09:14:36 crc kubenswrapper[4744]: I1008 09:14:36.646868 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:36 crc kubenswrapper[4744]: I1008 09:14:36.649856 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:14:37 crc kubenswrapper[4744]: I1008 09:14:37.575559 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:14:45 crc kubenswrapper[4744]: I1008 09:14:45.849104 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bw5th" Oct 08 09:14:47 crc kubenswrapper[4744]: E1008 09:14:47.435040 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 08 09:14:47 crc kubenswrapper[4744]: E1008 09:14:47.435587 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5gl2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-n7rwc_openshift-marketplace(baee305b-d6aa-41ac-a360-77ceb69f5db0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 09:14:47 crc kubenswrapper[4744]: E1008 09:14:47.436858 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-n7rwc" podUID="baee305b-d6aa-41ac-a360-77ceb69f5db0" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.741762 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-n7rwc" podUID="baee305b-d6aa-41ac-a360-77ceb69f5db0" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.898617 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.899837 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w5jx6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-pvc9m_openshift-marketplace(62e5fd77-7efc-455d-86d0-763bb040f2d0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.901653 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-pvc9m" podUID="62e5fd77-7efc-455d-86d0-763bb040f2d0" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.918362 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.918511 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zr262,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rsqj4_openshift-marketplace(d28208cc-1278-45d6-9d2e-a65a6d0c94d8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.919668 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rsqj4" podUID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.928036 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.928264 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s6jjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6qmwr_openshift-marketplace(432cd400-49d6-4dc1-a40d-182f5c6f4556): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.929850 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-6qmwr" podUID="432cd400-49d6-4dc1-a40d-182f5c6f4556" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.942259 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.945199 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2k2zn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-svt5h_openshift-marketplace(c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.948683 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-svt5h" podUID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.962507 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.962783 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6ssmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lnwmc_openshift-marketplace(e33828e9-5ea7-4f7d-ab35-474268801800): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 09:14:48 crc kubenswrapper[4744]: E1008 09:14:48.964048 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-lnwmc" podUID="e33828e9-5ea7-4f7d-ab35-474268801800" Oct 08 09:14:49 crc kubenswrapper[4744]: I1008 09:14:49.186197 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-cdcbz"] Oct 08 09:14:49 crc kubenswrapper[4744]: W1008 09:14:49.198907 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe364129_0f0b_41d4_b5f8_1b1def9a1c38.slice/crio-a7fdcef73dd33a8ed81fff7cf234a49c14634f89c9ab44bc1343b4722506ac78 WatchSource:0}: Error finding container a7fdcef73dd33a8ed81fff7cf234a49c14634f89c9ab44bc1343b4722506ac78: Status 404 returned error can't find the container with id a7fdcef73dd33a8ed81fff7cf234a49c14634f89c9ab44bc1343b4722506ac78 Oct 08 09:14:49 crc kubenswrapper[4744]: I1008 09:14:49.202070 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msvkb" event={"ID":"afc8f9d4-5df0-420c-9cba-aae825a33836","Type":"ContainerStarted","Data":"afe85c233e8657cd5557746720bd09bb712075ad17699dbdbb2d4a3d7708634d"} Oct 08 09:14:49 crc kubenswrapper[4744]: I1008 09:14:49.206530 4744 generic.go:334] "Generic (PLEG): container finished" podID="3c40f62a-9591-4187-aaf5-bbf7319d013c" containerID="7b2b5d94668a12b3883c14ad81f8afade08a21e0e39a403309ca7850d2435f49" exitCode=0 Oct 08 09:14:49 crc kubenswrapper[4744]: I1008 09:14:49.206625 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85kw7" event={"ID":"3c40f62a-9591-4187-aaf5-bbf7319d013c","Type":"ContainerDied","Data":"7b2b5d94668a12b3883c14ad81f8afade08a21e0e39a403309ca7850d2435f49"} Oct 08 09:14:49 crc kubenswrapper[4744]: E1008 09:14:49.212313 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lnwmc" podUID="e33828e9-5ea7-4f7d-ab35-474268801800" Oct 08 09:14:49 crc kubenswrapper[4744]: E1008 09:14:49.212420 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-pvc9m" podUID="62e5fd77-7efc-455d-86d0-763bb040f2d0" Oct 08 09:14:49 crc kubenswrapper[4744]: E1008 09:14:49.218724 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rsqj4" podUID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" Oct 08 09:14:49 crc kubenswrapper[4744]: E1008 09:14:49.218846 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-svt5h" podUID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" Oct 08 09:14:49 crc kubenswrapper[4744]: E1008 09:14:49.220562 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6qmwr" podUID="432cd400-49d6-4dc1-a40d-182f5c6f4556" Oct 08 09:14:49 crc kubenswrapper[4744]: I1008 09:14:49.690093 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:14:49 crc kubenswrapper[4744]: I1008 09:14:49.690814 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:14:50 crc kubenswrapper[4744]: I1008 09:14:50.218228 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" event={"ID":"be364129-0f0b-41d4-b5f8-1b1def9a1c38","Type":"ContainerStarted","Data":"4e3f9611faf649b1f51cbb5755dd43ac3f1b31aaf33492c6bbdb313dee48dc6f"} Oct 08 09:14:50 crc kubenswrapper[4744]: I1008 09:14:50.218997 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" event={"ID":"be364129-0f0b-41d4-b5f8-1b1def9a1c38","Type":"ContainerStarted","Data":"3358c99950799ffefc5db974549406e6f22cf63ac81f3769842653e56e19ac04"} Oct 08 09:14:50 crc kubenswrapper[4744]: I1008 09:14:50.219055 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-cdcbz" event={"ID":"be364129-0f0b-41d4-b5f8-1b1def9a1c38","Type":"ContainerStarted","Data":"a7fdcef73dd33a8ed81fff7cf234a49c14634f89c9ab44bc1343b4722506ac78"} Oct 08 09:14:50 crc kubenswrapper[4744]: I1008 09:14:50.222201 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85kw7" event={"ID":"3c40f62a-9591-4187-aaf5-bbf7319d013c","Type":"ContainerStarted","Data":"5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9"} Oct 08 09:14:50 crc kubenswrapper[4744]: I1008 09:14:50.227903 4744 generic.go:334] "Generic (PLEG): container finished" podID="afc8f9d4-5df0-420c-9cba-aae825a33836" containerID="afe85c233e8657cd5557746720bd09bb712075ad17699dbdbb2d4a3d7708634d" exitCode=0 Oct 08 09:14:50 crc kubenswrapper[4744]: I1008 09:14:50.227990 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msvkb" event={"ID":"afc8f9d4-5df0-420c-9cba-aae825a33836","Type":"ContainerDied","Data":"afe85c233e8657cd5557746720bd09bb712075ad17699dbdbb2d4a3d7708634d"} Oct 08 09:14:50 crc kubenswrapper[4744]: I1008 09:14:50.257717 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-cdcbz" podStartSLOduration=165.257684937 podStartE2EDuration="2m45.257684937s" podCreationTimestamp="2025-10-08 09:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:14:50.254581975 +0000 UTC m=+185.502227294" watchObservedRunningTime="2025-10-08 09:14:50.257684937 +0000 UTC m=+185.505330246" Oct 08 09:14:50 crc kubenswrapper[4744]: I1008 09:14:50.317794 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-85kw7" podStartSLOduration=4.829137562 podStartE2EDuration="37.317765157s" podCreationTimestamp="2025-10-08 09:14:13 +0000 UTC" firstStartedPulling="2025-10-08 09:14:17.128847955 +0000 UTC m=+152.376493194" lastFinishedPulling="2025-10-08 09:14:49.61747554 +0000 UTC m=+184.865120789" observedRunningTime="2025-10-08 09:14:50.289459888 +0000 UTC m=+185.537105157" watchObservedRunningTime="2025-10-08 09:14:50.317765157 +0000 UTC m=+185.565410406" Oct 08 09:14:51 crc kubenswrapper[4744]: I1008 09:14:51.241288 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msvkb" event={"ID":"afc8f9d4-5df0-420c-9cba-aae825a33836","Type":"ContainerStarted","Data":"6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a"} Oct 08 09:14:53 crc kubenswrapper[4744]: I1008 09:14:53.781218 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 09:14:53 crc kubenswrapper[4744]: I1008 09:14:53.800159 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-msvkb" podStartSLOduration=6.369403126 podStartE2EDuration="38.800137086s" podCreationTimestamp="2025-10-08 09:14:15 +0000 UTC" firstStartedPulling="2025-10-08 09:14:18.402710205 +0000 UTC m=+153.650355444" lastFinishedPulling="2025-10-08 09:14:50.833444145 +0000 UTC m=+186.081089404" observedRunningTime="2025-10-08 09:14:51.271956275 +0000 UTC m=+186.519601534" watchObservedRunningTime="2025-10-08 09:14:53.800137086 +0000 UTC m=+189.047782335" Oct 08 09:14:54 crc kubenswrapper[4744]: I1008 09:14:54.388730 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:54 crc kubenswrapper[4744]: I1008 09:14:54.388819 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:54 crc kubenswrapper[4744]: I1008 09:14:54.707706 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:55 crc kubenswrapper[4744]: I1008 09:14:55.296095 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:14:55 crc kubenswrapper[4744]: I1008 09:14:55.448000 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:55 crc kubenswrapper[4744]: I1008 09:14:55.448064 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:14:56 crc kubenswrapper[4744]: I1008 09:14:56.490842 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-msvkb" podUID="afc8f9d4-5df0-420c-9cba-aae825a33836" containerName="registry-server" probeResult="failure" output=< Oct 08 09:14:56 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 09:14:56 crc kubenswrapper[4744]: > Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.158277 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2"] Oct 08 09:15:00 crc kubenswrapper[4744]: E1008 09:15:00.161740 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bacbc855-fe99-4250-9da3-b64e95ac7c67" containerName="collect-profiles" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.161906 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bacbc855-fe99-4250-9da3-b64e95ac7c67" containerName="collect-profiles" Oct 08 09:15:00 crc kubenswrapper[4744]: E1008 09:15:00.162052 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c81e8bf0-59bf-41ff-a992-7bd3fed6a970" containerName="pruner" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.162185 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c81e8bf0-59bf-41ff-a992-7bd3fed6a970" containerName="pruner" Oct 08 09:15:00 crc kubenswrapper[4744]: E1008 09:15:00.162316 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2796b5f-00d8-491e-b4b0-e5ebd034674b" containerName="pruner" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.162484 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2796b5f-00d8-491e-b4b0-e5ebd034674b" containerName="pruner" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.162865 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2796b5f-00d8-491e-b4b0-e5ebd034674b" containerName="pruner" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.163054 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bacbc855-fe99-4250-9da3-b64e95ac7c67" containerName="collect-profiles" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.163205 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c81e8bf0-59bf-41ff-a992-7bd3fed6a970" containerName="pruner" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.164111 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.168523 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.172179 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.174009 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2"] Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.267402 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c847191-fb16-4cdc-8b54-ee09e3571cce-config-volume\") pod \"collect-profiles-29331915-7gnz2\" (UID: \"9c847191-fb16-4cdc-8b54-ee09e3571cce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.267477 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c847191-fb16-4cdc-8b54-ee09e3571cce-secret-volume\") pod \"collect-profiles-29331915-7gnz2\" (UID: \"9c847191-fb16-4cdc-8b54-ee09e3571cce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.267646 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52t6b\" (UniqueName: \"kubernetes.io/projected/9c847191-fb16-4cdc-8b54-ee09e3571cce-kube-api-access-52t6b\") pod \"collect-profiles-29331915-7gnz2\" (UID: \"9c847191-fb16-4cdc-8b54-ee09e3571cce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.369282 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c847191-fb16-4cdc-8b54-ee09e3571cce-config-volume\") pod \"collect-profiles-29331915-7gnz2\" (UID: \"9c847191-fb16-4cdc-8b54-ee09e3571cce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.369694 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c847191-fb16-4cdc-8b54-ee09e3571cce-secret-volume\") pod \"collect-profiles-29331915-7gnz2\" (UID: \"9c847191-fb16-4cdc-8b54-ee09e3571cce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.369923 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52t6b\" (UniqueName: \"kubernetes.io/projected/9c847191-fb16-4cdc-8b54-ee09e3571cce-kube-api-access-52t6b\") pod \"collect-profiles-29331915-7gnz2\" (UID: \"9c847191-fb16-4cdc-8b54-ee09e3571cce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.374005 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c847191-fb16-4cdc-8b54-ee09e3571cce-config-volume\") pod \"collect-profiles-29331915-7gnz2\" (UID: \"9c847191-fb16-4cdc-8b54-ee09e3571cce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.380032 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c847191-fb16-4cdc-8b54-ee09e3571cce-secret-volume\") pod \"collect-profiles-29331915-7gnz2\" (UID: \"9c847191-fb16-4cdc-8b54-ee09e3571cce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.390532 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52t6b\" (UniqueName: \"kubernetes.io/projected/9c847191-fb16-4cdc-8b54-ee09e3571cce-kube-api-access-52t6b\") pod \"collect-profiles-29331915-7gnz2\" (UID: \"9c847191-fb16-4cdc-8b54-ee09e3571cce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.497243 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:00 crc kubenswrapper[4744]: I1008 09:15:00.947166 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2"] Oct 08 09:15:00 crc kubenswrapper[4744]: W1008 09:15:00.959863 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c847191_fb16_4cdc_8b54_ee09e3571cce.slice/crio-dd67bb2f3421ca24786f7d7b0461d945b67f9eb0ee9065e07b343f45ff08177e WatchSource:0}: Error finding container dd67bb2f3421ca24786f7d7b0461d945b67f9eb0ee9065e07b343f45ff08177e: Status 404 returned error can't find the container with id dd67bb2f3421ca24786f7d7b0461d945b67f9eb0ee9065e07b343f45ff08177e Oct 08 09:15:01 crc kubenswrapper[4744]: I1008 09:15:01.303608 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" event={"ID":"9c847191-fb16-4cdc-8b54-ee09e3571cce","Type":"ContainerStarted","Data":"dd67bb2f3421ca24786f7d7b0461d945b67f9eb0ee9065e07b343f45ff08177e"} Oct 08 09:15:02 crc kubenswrapper[4744]: I1008 09:15:02.310454 4744 generic.go:334] "Generic (PLEG): container finished" podID="9c847191-fb16-4cdc-8b54-ee09e3571cce" containerID="37a4f12941e6d9f3d3fef029d1a710111f71d8be62a0c60ed6e27260a915b35f" exitCode=0 Oct 08 09:15:02 crc kubenswrapper[4744]: I1008 09:15:02.310566 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" event={"ID":"9c847191-fb16-4cdc-8b54-ee09e3571cce","Type":"ContainerDied","Data":"37a4f12941e6d9f3d3fef029d1a710111f71d8be62a0c60ed6e27260a915b35f"} Oct 08 09:15:03 crc kubenswrapper[4744]: I1008 09:15:03.701202 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:03 crc kubenswrapper[4744]: I1008 09:15:03.825034 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c847191-fb16-4cdc-8b54-ee09e3571cce-config-volume\") pod \"9c847191-fb16-4cdc-8b54-ee09e3571cce\" (UID: \"9c847191-fb16-4cdc-8b54-ee09e3571cce\") " Oct 08 09:15:03 crc kubenswrapper[4744]: I1008 09:15:03.825137 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52t6b\" (UniqueName: \"kubernetes.io/projected/9c847191-fb16-4cdc-8b54-ee09e3571cce-kube-api-access-52t6b\") pod \"9c847191-fb16-4cdc-8b54-ee09e3571cce\" (UID: \"9c847191-fb16-4cdc-8b54-ee09e3571cce\") " Oct 08 09:15:03 crc kubenswrapper[4744]: I1008 09:15:03.825178 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c847191-fb16-4cdc-8b54-ee09e3571cce-secret-volume\") pod \"9c847191-fb16-4cdc-8b54-ee09e3571cce\" (UID: \"9c847191-fb16-4cdc-8b54-ee09e3571cce\") " Oct 08 09:15:03 crc kubenswrapper[4744]: I1008 09:15:03.826001 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c847191-fb16-4cdc-8b54-ee09e3571cce-config-volume" (OuterVolumeSpecName: "config-volume") pod "9c847191-fb16-4cdc-8b54-ee09e3571cce" (UID: "9c847191-fb16-4cdc-8b54-ee09e3571cce"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:15:03 crc kubenswrapper[4744]: I1008 09:15:03.831675 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c847191-fb16-4cdc-8b54-ee09e3571cce-kube-api-access-52t6b" (OuterVolumeSpecName: "kube-api-access-52t6b") pod "9c847191-fb16-4cdc-8b54-ee09e3571cce" (UID: "9c847191-fb16-4cdc-8b54-ee09e3571cce"). InnerVolumeSpecName "kube-api-access-52t6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:15:03 crc kubenswrapper[4744]: I1008 09:15:03.832178 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c847191-fb16-4cdc-8b54-ee09e3571cce-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9c847191-fb16-4cdc-8b54-ee09e3571cce" (UID: "9c847191-fb16-4cdc-8b54-ee09e3571cce"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:15:03 crc kubenswrapper[4744]: I1008 09:15:03.931711 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52t6b\" (UniqueName: \"kubernetes.io/projected/9c847191-fb16-4cdc-8b54-ee09e3571cce-kube-api-access-52t6b\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:03 crc kubenswrapper[4744]: I1008 09:15:03.932114 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9c847191-fb16-4cdc-8b54-ee09e3571cce-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:03 crc kubenswrapper[4744]: I1008 09:15:03.932128 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9c847191-fb16-4cdc-8b54-ee09e3571cce-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:04 crc kubenswrapper[4744]: I1008 09:15:04.326004 4744 generic.go:334] "Generic (PLEG): container finished" podID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" containerID="d1304c2c5cf4962ce5d69fd4e8492821e33636b7574545b48c877e49e8850060" exitCode=0 Oct 08 09:15:04 crc kubenswrapper[4744]: I1008 09:15:04.326081 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsqj4" event={"ID":"d28208cc-1278-45d6-9d2e-a65a6d0c94d8","Type":"ContainerDied","Data":"d1304c2c5cf4962ce5d69fd4e8492821e33636b7574545b48c877e49e8850060"} Oct 08 09:15:04 crc kubenswrapper[4744]: I1008 09:15:04.339268 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7rwc" event={"ID":"baee305b-d6aa-41ac-a360-77ceb69f5db0","Type":"ContainerStarted","Data":"8e3380ad2b764a851403a900b5b67bc0bfce8ec5b8d3c950aa0dbb5e8933b3b0"} Oct 08 09:15:04 crc kubenswrapper[4744]: I1008 09:15:04.350776 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" event={"ID":"9c847191-fb16-4cdc-8b54-ee09e3571cce","Type":"ContainerDied","Data":"dd67bb2f3421ca24786f7d7b0461d945b67f9eb0ee9065e07b343f45ff08177e"} Oct 08 09:15:04 crc kubenswrapper[4744]: I1008 09:15:04.350823 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd67bb2f3421ca24786f7d7b0461d945b67f9eb0ee9065e07b343f45ff08177e" Oct 08 09:15:04 crc kubenswrapper[4744]: I1008 09:15:04.350893 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2" Oct 08 09:15:04 crc kubenswrapper[4744]: I1008 09:15:04.355051 4744 generic.go:334] "Generic (PLEG): container finished" podID="e33828e9-5ea7-4f7d-ab35-474268801800" containerID="aea4f5d8c68ab26809b8ddd65d4bd863accca34ebf16a99f2b3cc9d546edf71b" exitCode=0 Oct 08 09:15:04 crc kubenswrapper[4744]: I1008 09:15:04.355123 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnwmc" event={"ID":"e33828e9-5ea7-4f7d-ab35-474268801800","Type":"ContainerDied","Data":"aea4f5d8c68ab26809b8ddd65d4bd863accca34ebf16a99f2b3cc9d546edf71b"} Oct 08 09:15:04 crc kubenswrapper[4744]: I1008 09:15:04.358614 4744 generic.go:334] "Generic (PLEG): container finished" podID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" containerID="602a06c61efbdf18f91dc32c2e37e819b61123ec4be8b2f10e352104d62799db" exitCode=0 Oct 08 09:15:04 crc kubenswrapper[4744]: I1008 09:15:04.358669 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svt5h" event={"ID":"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc","Type":"ContainerDied","Data":"602a06c61efbdf18f91dc32c2e37e819b61123ec4be8b2f10e352104d62799db"} Oct 08 09:15:04 crc kubenswrapper[4744]: I1008 09:15:04.363074 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qmwr" event={"ID":"432cd400-49d6-4dc1-a40d-182f5c6f4556","Type":"ContainerStarted","Data":"0488b56ca4ce620970a8f3b16fef6a6a7dcc90dd757bed319d90e6f938e08cb4"} Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.371112 4744 generic.go:334] "Generic (PLEG): container finished" podID="baee305b-d6aa-41ac-a360-77ceb69f5db0" containerID="8e3380ad2b764a851403a900b5b67bc0bfce8ec5b8d3c950aa0dbb5e8933b3b0" exitCode=0 Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.371187 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7rwc" event={"ID":"baee305b-d6aa-41ac-a360-77ceb69f5db0","Type":"ContainerDied","Data":"8e3380ad2b764a851403a900b5b67bc0bfce8ec5b8d3c950aa0dbb5e8933b3b0"} Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.376250 4744 generic.go:334] "Generic (PLEG): container finished" podID="62e5fd77-7efc-455d-86d0-763bb040f2d0" containerID="a07b9b00942b4a4f84cb103c3bb84780bf39d126b98e0f6e3a803c9cca73e381" exitCode=0 Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.376318 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvc9m" event={"ID":"62e5fd77-7efc-455d-86d0-763bb040f2d0","Type":"ContainerDied","Data":"a07b9b00942b4a4f84cb103c3bb84780bf39d126b98e0f6e3a803c9cca73e381"} Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.380153 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnwmc" event={"ID":"e33828e9-5ea7-4f7d-ab35-474268801800","Type":"ContainerStarted","Data":"5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6"} Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.387039 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svt5h" event={"ID":"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc","Type":"ContainerStarted","Data":"d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897"} Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.389638 4744 generic.go:334] "Generic (PLEG): container finished" podID="432cd400-49d6-4dc1-a40d-182f5c6f4556" containerID="0488b56ca4ce620970a8f3b16fef6a6a7dcc90dd757bed319d90e6f938e08cb4" exitCode=0 Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.389696 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qmwr" event={"ID":"432cd400-49d6-4dc1-a40d-182f5c6f4556","Type":"ContainerDied","Data":"0488b56ca4ce620970a8f3b16fef6a6a7dcc90dd757bed319d90e6f938e08cb4"} Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.403551 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsqj4" event={"ID":"d28208cc-1278-45d6-9d2e-a65a6d0c94d8","Type":"ContainerStarted","Data":"a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf"} Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.442500 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lnwmc" podStartSLOduration=5.291117113 podStartE2EDuration="54.442481416s" podCreationTimestamp="2025-10-08 09:14:11 +0000 UTC" firstStartedPulling="2025-10-08 09:14:15.732489947 +0000 UTC m=+150.980135186" lastFinishedPulling="2025-10-08 09:15:04.88385424 +0000 UTC m=+200.131499489" observedRunningTime="2025-10-08 09:15:05.441577987 +0000 UTC m=+200.689223236" watchObservedRunningTime="2025-10-08 09:15:05.442481416 +0000 UTC m=+200.690126655" Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.505569 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-svt5h" podStartSLOduration=5.319219141 podStartE2EDuration="51.505551763s" podCreationTimestamp="2025-10-08 09:14:14 +0000 UTC" firstStartedPulling="2025-10-08 09:14:18.639948324 +0000 UTC m=+153.887593563" lastFinishedPulling="2025-10-08 09:15:04.826280946 +0000 UTC m=+200.073926185" observedRunningTime="2025-10-08 09:15:05.501514191 +0000 UTC m=+200.749159440" watchObservedRunningTime="2025-10-08 09:15:05.505551763 +0000 UTC m=+200.753196992" Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.512471 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.582711 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rsqj4" podStartSLOduration=5.162293624 podStartE2EDuration="54.582681461s" podCreationTimestamp="2025-10-08 09:14:11 +0000 UTC" firstStartedPulling="2025-10-08 09:14:15.76800007 +0000 UTC m=+151.015645309" lastFinishedPulling="2025-10-08 09:15:05.188387907 +0000 UTC m=+200.436033146" observedRunningTime="2025-10-08 09:15:05.539051033 +0000 UTC m=+200.786696282" watchObservedRunningTime="2025-10-08 09:15:05.582681461 +0000 UTC m=+200.830326700" Oct 08 09:15:05 crc kubenswrapper[4744]: I1008 09:15:05.584317 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:15:06 crc kubenswrapper[4744]: I1008 09:15:06.413831 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qmwr" event={"ID":"432cd400-49d6-4dc1-a40d-182f5c6f4556","Type":"ContainerStarted","Data":"043047350a72aeef1dc3d32608103d9ffa7338b8089f926e6e881bd8ae724e14"} Oct 08 09:15:06 crc kubenswrapper[4744]: I1008 09:15:06.416965 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7rwc" event={"ID":"baee305b-d6aa-41ac-a360-77ceb69f5db0","Type":"ContainerStarted","Data":"a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27"} Oct 08 09:15:06 crc kubenswrapper[4744]: I1008 09:15:06.420052 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvc9m" event={"ID":"62e5fd77-7efc-455d-86d0-763bb040f2d0","Type":"ContainerStarted","Data":"18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8"} Oct 08 09:15:06 crc kubenswrapper[4744]: I1008 09:15:06.439809 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6qmwr" podStartSLOduration=5.192633734 podStartE2EDuration="52.439785148s" podCreationTimestamp="2025-10-08 09:14:14 +0000 UTC" firstStartedPulling="2025-10-08 09:14:18.639594503 +0000 UTC m=+153.887239742" lastFinishedPulling="2025-10-08 09:15:05.886745917 +0000 UTC m=+201.134391156" observedRunningTime="2025-10-08 09:15:06.435315976 +0000 UTC m=+201.682961235" watchObservedRunningTime="2025-10-08 09:15:06.439785148 +0000 UTC m=+201.687430377" Oct 08 09:15:06 crc kubenswrapper[4744]: I1008 09:15:06.493401 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n7rwc" podStartSLOduration=5.638643213 podStartE2EDuration="54.493379881s" podCreationTimestamp="2025-10-08 09:14:12 +0000 UTC" firstStartedPulling="2025-10-08 09:14:17.168292364 +0000 UTC m=+152.415937603" lastFinishedPulling="2025-10-08 09:15:06.023029042 +0000 UTC m=+201.270674271" observedRunningTime="2025-10-08 09:15:06.48898976 +0000 UTC m=+201.736635019" watchObservedRunningTime="2025-10-08 09:15:06.493379881 +0000 UTC m=+201.741025120" Oct 08 09:15:06 crc kubenswrapper[4744]: I1008 09:15:06.515115 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pvc9m" podStartSLOduration=5.529049003 podStartE2EDuration="55.515092987s" podCreationTimestamp="2025-10-08 09:14:11 +0000 UTC" firstStartedPulling="2025-10-08 09:14:15.82473615 +0000 UTC m=+151.072381389" lastFinishedPulling="2025-10-08 09:15:05.810780134 +0000 UTC m=+201.058425373" observedRunningTime="2025-10-08 09:15:06.510038544 +0000 UTC m=+201.757683793" watchObservedRunningTime="2025-10-08 09:15:06.515092987 +0000 UTC m=+201.762738226" Oct 08 09:15:08 crc kubenswrapper[4744]: I1008 09:15:08.498492 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-msvkb"] Oct 08 09:15:08 crc kubenswrapper[4744]: I1008 09:15:08.499592 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-msvkb" podUID="afc8f9d4-5df0-420c-9cba-aae825a33836" containerName="registry-server" containerID="cri-o://6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a" gracePeriod=2 Oct 08 09:15:08 crc kubenswrapper[4744]: I1008 09:15:08.997623 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.116933 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs2lc\" (UniqueName: \"kubernetes.io/projected/afc8f9d4-5df0-420c-9cba-aae825a33836-kube-api-access-gs2lc\") pod \"afc8f9d4-5df0-420c-9cba-aae825a33836\" (UID: \"afc8f9d4-5df0-420c-9cba-aae825a33836\") " Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.117566 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8f9d4-5df0-420c-9cba-aae825a33836-catalog-content\") pod \"afc8f9d4-5df0-420c-9cba-aae825a33836\" (UID: \"afc8f9d4-5df0-420c-9cba-aae825a33836\") " Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.117618 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8f9d4-5df0-420c-9cba-aae825a33836-utilities\") pod \"afc8f9d4-5df0-420c-9cba-aae825a33836\" (UID: \"afc8f9d4-5df0-420c-9cba-aae825a33836\") " Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.118769 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc8f9d4-5df0-420c-9cba-aae825a33836-utilities" (OuterVolumeSpecName: "utilities") pod "afc8f9d4-5df0-420c-9cba-aae825a33836" (UID: "afc8f9d4-5df0-420c-9cba-aae825a33836"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.128441 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc8f9d4-5df0-420c-9cba-aae825a33836-kube-api-access-gs2lc" (OuterVolumeSpecName: "kube-api-access-gs2lc") pod "afc8f9d4-5df0-420c-9cba-aae825a33836" (UID: "afc8f9d4-5df0-420c-9cba-aae825a33836"). InnerVolumeSpecName "kube-api-access-gs2lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.220777 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs2lc\" (UniqueName: \"kubernetes.io/projected/afc8f9d4-5df0-420c-9cba-aae825a33836-kube-api-access-gs2lc\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.220850 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8f9d4-5df0-420c-9cba-aae825a33836-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.227424 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc8f9d4-5df0-420c-9cba-aae825a33836-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afc8f9d4-5df0-420c-9cba-aae825a33836" (UID: "afc8f9d4-5df0-420c-9cba-aae825a33836"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.323527 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8f9d4-5df0-420c-9cba-aae825a33836-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.461046 4744 generic.go:334] "Generic (PLEG): container finished" podID="afc8f9d4-5df0-420c-9cba-aae825a33836" containerID="6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a" exitCode=0 Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.461169 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-msvkb" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.462611 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msvkb" event={"ID":"afc8f9d4-5df0-420c-9cba-aae825a33836","Type":"ContainerDied","Data":"6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a"} Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.462683 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-msvkb" event={"ID":"afc8f9d4-5df0-420c-9cba-aae825a33836","Type":"ContainerDied","Data":"6c6090ee653702b139b3ae833f54a6208642f6a2ad437f7b12f4899995c53324"} Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.462716 4744 scope.go:117] "RemoveContainer" containerID="6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.496970 4744 scope.go:117] "RemoveContainer" containerID="afe85c233e8657cd5557746720bd09bb712075ad17699dbdbb2d4a3d7708634d" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.509796 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-msvkb"] Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.519567 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-msvkb"] Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.529780 4744 scope.go:117] "RemoveContainer" containerID="fe0f4abdf5948e21dbb672e73f63fd9f930a78dfa3081b67d5b876e3441eaab0" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.557614 4744 scope.go:117] "RemoveContainer" containerID="6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a" Oct 08 09:15:09 crc kubenswrapper[4744]: E1008 09:15:09.558270 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a\": container with ID starting with 6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a not found: ID does not exist" containerID="6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.558328 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a"} err="failed to get container status \"6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a\": rpc error: code = NotFound desc = could not find container \"6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a\": container with ID starting with 6bcf17111df7e37931ac1f432554100109166b885395deb03d417fdc8208862a not found: ID does not exist" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.558399 4744 scope.go:117] "RemoveContainer" containerID="afe85c233e8657cd5557746720bd09bb712075ad17699dbdbb2d4a3d7708634d" Oct 08 09:15:09 crc kubenswrapper[4744]: E1008 09:15:09.559448 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afe85c233e8657cd5557746720bd09bb712075ad17699dbdbb2d4a3d7708634d\": container with ID starting with afe85c233e8657cd5557746720bd09bb712075ad17699dbdbb2d4a3d7708634d not found: ID does not exist" containerID="afe85c233e8657cd5557746720bd09bb712075ad17699dbdbb2d4a3d7708634d" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.559474 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afe85c233e8657cd5557746720bd09bb712075ad17699dbdbb2d4a3d7708634d"} err="failed to get container status \"afe85c233e8657cd5557746720bd09bb712075ad17699dbdbb2d4a3d7708634d\": rpc error: code = NotFound desc = could not find container \"afe85c233e8657cd5557746720bd09bb712075ad17699dbdbb2d4a3d7708634d\": container with ID starting with afe85c233e8657cd5557746720bd09bb712075ad17699dbdbb2d4a3d7708634d not found: ID does not exist" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.559492 4744 scope.go:117] "RemoveContainer" containerID="fe0f4abdf5948e21dbb672e73f63fd9f930a78dfa3081b67d5b876e3441eaab0" Oct 08 09:15:09 crc kubenswrapper[4744]: E1008 09:15:09.559936 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe0f4abdf5948e21dbb672e73f63fd9f930a78dfa3081b67d5b876e3441eaab0\": container with ID starting with fe0f4abdf5948e21dbb672e73f63fd9f930a78dfa3081b67d5b876e3441eaab0 not found: ID does not exist" containerID="fe0f4abdf5948e21dbb672e73f63fd9f930a78dfa3081b67d5b876e3441eaab0" Oct 08 09:15:09 crc kubenswrapper[4744]: I1008 09:15:09.560003 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe0f4abdf5948e21dbb672e73f63fd9f930a78dfa3081b67d5b876e3441eaab0"} err="failed to get container status \"fe0f4abdf5948e21dbb672e73f63fd9f930a78dfa3081b67d5b876e3441eaab0\": rpc error: code = NotFound desc = could not find container \"fe0f4abdf5948e21dbb672e73f63fd9f930a78dfa3081b67d5b876e3441eaab0\": container with ID starting with fe0f4abdf5948e21dbb672e73f63fd9f930a78dfa3081b67d5b876e3441eaab0 not found: ID does not exist" Oct 08 09:15:11 crc kubenswrapper[4744]: I1008 09:15:11.491102 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc8f9d4-5df0-420c-9cba-aae825a33836" path="/var/lib/kubelet/pods/afc8f9d4-5df0-420c-9cba-aae825a33836/volumes" Oct 08 09:15:11 crc kubenswrapper[4744]: I1008 09:15:11.494044 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x97sx"] Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.139975 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.140593 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.189744 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.267194 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.267269 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.323761 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.548994 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.591248 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.893858 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lnwmc"] Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.992046 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.992943 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.996767 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:15:12 crc kubenswrapper[4744]: I1008 09:15:12.996840 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:15:13 crc kubenswrapper[4744]: I1008 09:15:13.039418 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:15:13 crc kubenswrapper[4744]: I1008 09:15:13.043136 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:15:13 crc kubenswrapper[4744]: I1008 09:15:13.563855 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:15:13 crc kubenswrapper[4744]: I1008 09:15:13.585918 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:15:14 crc kubenswrapper[4744]: I1008 09:15:14.515320 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lnwmc" podUID="e33828e9-5ea7-4f7d-ab35-474268801800" containerName="registry-server" containerID="cri-o://5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6" gracePeriod=2 Oct 08 09:15:14 crc kubenswrapper[4744]: I1008 09:15:14.784427 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:15:14 crc kubenswrapper[4744]: I1008 09:15:14.784472 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:15:14 crc kubenswrapper[4744]: I1008 09:15:14.845846 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:15:14 crc kubenswrapper[4744]: I1008 09:15:14.916319 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.020786 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e33828e9-5ea7-4f7d-ab35-474268801800-utilities\") pod \"e33828e9-5ea7-4f7d-ab35-474268801800\" (UID: \"e33828e9-5ea7-4f7d-ab35-474268801800\") " Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.020904 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e33828e9-5ea7-4f7d-ab35-474268801800-catalog-content\") pod \"e33828e9-5ea7-4f7d-ab35-474268801800\" (UID: \"e33828e9-5ea7-4f7d-ab35-474268801800\") " Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.020943 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ssmg\" (UniqueName: \"kubernetes.io/projected/e33828e9-5ea7-4f7d-ab35-474268801800-kube-api-access-6ssmg\") pod \"e33828e9-5ea7-4f7d-ab35-474268801800\" (UID: \"e33828e9-5ea7-4f7d-ab35-474268801800\") " Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.023069 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e33828e9-5ea7-4f7d-ab35-474268801800-utilities" (OuterVolumeSpecName: "utilities") pod "e33828e9-5ea7-4f7d-ab35-474268801800" (UID: "e33828e9-5ea7-4f7d-ab35-474268801800"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.026472 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e33828e9-5ea7-4f7d-ab35-474268801800-kube-api-access-6ssmg" (OuterVolumeSpecName: "kube-api-access-6ssmg") pod "e33828e9-5ea7-4f7d-ab35-474268801800" (UID: "e33828e9-5ea7-4f7d-ab35-474268801800"). InnerVolumeSpecName "kube-api-access-6ssmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.084009 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e33828e9-5ea7-4f7d-ab35-474268801800-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e33828e9-5ea7-4f7d-ab35-474268801800" (UID: "e33828e9-5ea7-4f7d-ab35-474268801800"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.122193 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e33828e9-5ea7-4f7d-ab35-474268801800-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.122228 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e33828e9-5ea7-4f7d-ab35-474268801800-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.122244 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ssmg\" (UniqueName: \"kubernetes.io/projected/e33828e9-5ea7-4f7d-ab35-474268801800-kube-api-access-6ssmg\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.136736 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.136799 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.184447 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.298867 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n7rwc"] Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.527102 4744 generic.go:334] "Generic (PLEG): container finished" podID="e33828e9-5ea7-4f7d-ab35-474268801800" containerID="5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6" exitCode=0 Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.527223 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lnwmc" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.527311 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnwmc" event={"ID":"e33828e9-5ea7-4f7d-ab35-474268801800","Type":"ContainerDied","Data":"5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6"} Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.527404 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lnwmc" event={"ID":"e33828e9-5ea7-4f7d-ab35-474268801800","Type":"ContainerDied","Data":"ddf39f5a5a0cae0279a98053d73f57c95861e93fce85192eb4bea007cf971db7"} Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.527433 4744 scope.go:117] "RemoveContainer" containerID="5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.528624 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n7rwc" podUID="baee305b-d6aa-41ac-a360-77ceb69f5db0" containerName="registry-server" containerID="cri-o://a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27" gracePeriod=2 Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.559512 4744 scope.go:117] "RemoveContainer" containerID="aea4f5d8c68ab26809b8ddd65d4bd863accca34ebf16a99f2b3cc9d546edf71b" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.559749 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lnwmc"] Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.561232 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lnwmc"] Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.596870 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.597144 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.600038 4744 scope.go:117] "RemoveContainer" containerID="b700c2f160368aa32f40a35c20835c669dd389a63912f2372c2f65a62e169352" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.646623 4744 scope.go:117] "RemoveContainer" containerID="5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6" Oct 08 09:15:15 crc kubenswrapper[4744]: E1008 09:15:15.647213 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6\": container with ID starting with 5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6 not found: ID does not exist" containerID="5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.647250 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6"} err="failed to get container status \"5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6\": rpc error: code = NotFound desc = could not find container \"5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6\": container with ID starting with 5725c1db29dae2078c28ce36defc83067d595f81b003eba10dbff43bca41bef6 not found: ID does not exist" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.647301 4744 scope.go:117] "RemoveContainer" containerID="aea4f5d8c68ab26809b8ddd65d4bd863accca34ebf16a99f2b3cc9d546edf71b" Oct 08 09:15:15 crc kubenswrapper[4744]: E1008 09:15:15.649021 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aea4f5d8c68ab26809b8ddd65d4bd863accca34ebf16a99f2b3cc9d546edf71b\": container with ID starting with aea4f5d8c68ab26809b8ddd65d4bd863accca34ebf16a99f2b3cc9d546edf71b not found: ID does not exist" containerID="aea4f5d8c68ab26809b8ddd65d4bd863accca34ebf16a99f2b3cc9d546edf71b" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.649056 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aea4f5d8c68ab26809b8ddd65d4bd863accca34ebf16a99f2b3cc9d546edf71b"} err="failed to get container status \"aea4f5d8c68ab26809b8ddd65d4bd863accca34ebf16a99f2b3cc9d546edf71b\": rpc error: code = NotFound desc = could not find container \"aea4f5d8c68ab26809b8ddd65d4bd863accca34ebf16a99f2b3cc9d546edf71b\": container with ID starting with aea4f5d8c68ab26809b8ddd65d4bd863accca34ebf16a99f2b3cc9d546edf71b not found: ID does not exist" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.649072 4744 scope.go:117] "RemoveContainer" containerID="b700c2f160368aa32f40a35c20835c669dd389a63912f2372c2f65a62e169352" Oct 08 09:15:15 crc kubenswrapper[4744]: E1008 09:15:15.649604 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b700c2f160368aa32f40a35c20835c669dd389a63912f2372c2f65a62e169352\": container with ID starting with b700c2f160368aa32f40a35c20835c669dd389a63912f2372c2f65a62e169352 not found: ID does not exist" containerID="b700c2f160368aa32f40a35c20835c669dd389a63912f2372c2f65a62e169352" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.649626 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b700c2f160368aa32f40a35c20835c669dd389a63912f2372c2f65a62e169352"} err="failed to get container status \"b700c2f160368aa32f40a35c20835c669dd389a63912f2372c2f65a62e169352\": rpc error: code = NotFound desc = could not find container \"b700c2f160368aa32f40a35c20835c669dd389a63912f2372c2f65a62e169352\": container with ID starting with b700c2f160368aa32f40a35c20835c669dd389a63912f2372c2f65a62e169352 not found: ID does not exist" Oct 08 09:15:15 crc kubenswrapper[4744]: I1008 09:15:15.962598 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.035728 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baee305b-d6aa-41ac-a360-77ceb69f5db0-utilities\") pod \"baee305b-d6aa-41ac-a360-77ceb69f5db0\" (UID: \"baee305b-d6aa-41ac-a360-77ceb69f5db0\") " Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.036130 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baee305b-d6aa-41ac-a360-77ceb69f5db0-catalog-content\") pod \"baee305b-d6aa-41ac-a360-77ceb69f5db0\" (UID: \"baee305b-d6aa-41ac-a360-77ceb69f5db0\") " Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.036212 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gl2n\" (UniqueName: \"kubernetes.io/projected/baee305b-d6aa-41ac-a360-77ceb69f5db0-kube-api-access-5gl2n\") pod \"baee305b-d6aa-41ac-a360-77ceb69f5db0\" (UID: \"baee305b-d6aa-41ac-a360-77ceb69f5db0\") " Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.037006 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baee305b-d6aa-41ac-a360-77ceb69f5db0-utilities" (OuterVolumeSpecName: "utilities") pod "baee305b-d6aa-41ac-a360-77ceb69f5db0" (UID: "baee305b-d6aa-41ac-a360-77ceb69f5db0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.048537 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baee305b-d6aa-41ac-a360-77ceb69f5db0-kube-api-access-5gl2n" (OuterVolumeSpecName: "kube-api-access-5gl2n") pod "baee305b-d6aa-41ac-a360-77ceb69f5db0" (UID: "baee305b-d6aa-41ac-a360-77ceb69f5db0"). InnerVolumeSpecName "kube-api-access-5gl2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.086966 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baee305b-d6aa-41ac-a360-77ceb69f5db0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "baee305b-d6aa-41ac-a360-77ceb69f5db0" (UID: "baee305b-d6aa-41ac-a360-77ceb69f5db0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.137107 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baee305b-d6aa-41ac-a360-77ceb69f5db0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.137152 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gl2n\" (UniqueName: \"kubernetes.io/projected/baee305b-d6aa-41ac-a360-77ceb69f5db0-kube-api-access-5gl2n\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.137161 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baee305b-d6aa-41ac-a360-77ceb69f5db0-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.535617 4744 generic.go:334] "Generic (PLEG): container finished" podID="baee305b-d6aa-41ac-a360-77ceb69f5db0" containerID="a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27" exitCode=0 Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.536518 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n7rwc" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.538268 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7rwc" event={"ID":"baee305b-d6aa-41ac-a360-77ceb69f5db0","Type":"ContainerDied","Data":"a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27"} Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.538339 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n7rwc" event={"ID":"baee305b-d6aa-41ac-a360-77ceb69f5db0","Type":"ContainerDied","Data":"ee28e3b3c0fc42e24fcc1e26d994f3c28fa89614813ab1fd9f7c458a54c6501b"} Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.538400 4744 scope.go:117] "RemoveContainer" containerID="a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.556597 4744 scope.go:117] "RemoveContainer" containerID="8e3380ad2b764a851403a900b5b67bc0bfce8ec5b8d3c950aa0dbb5e8933b3b0" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.575872 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n7rwc"] Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.580438 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n7rwc"] Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.585976 4744 scope.go:117] "RemoveContainer" containerID="9c3ae761e81730a0da4906c9dc39fb77f17c2ca8027ca2a411bd2340fabc141b" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.613856 4744 scope.go:117] "RemoveContainer" containerID="a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27" Oct 08 09:15:16 crc kubenswrapper[4744]: E1008 09:15:16.614460 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27\": container with ID starting with a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27 not found: ID does not exist" containerID="a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.614515 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27"} err="failed to get container status \"a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27\": rpc error: code = NotFound desc = could not find container \"a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27\": container with ID starting with a15aac27e9b6f68e79600d46194dde297dde1601d174ec854b39423b0a292a27 not found: ID does not exist" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.614546 4744 scope.go:117] "RemoveContainer" containerID="8e3380ad2b764a851403a900b5b67bc0bfce8ec5b8d3c950aa0dbb5e8933b3b0" Oct 08 09:15:16 crc kubenswrapper[4744]: E1008 09:15:16.615189 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e3380ad2b764a851403a900b5b67bc0bfce8ec5b8d3c950aa0dbb5e8933b3b0\": container with ID starting with 8e3380ad2b764a851403a900b5b67bc0bfce8ec5b8d3c950aa0dbb5e8933b3b0 not found: ID does not exist" containerID="8e3380ad2b764a851403a900b5b67bc0bfce8ec5b8d3c950aa0dbb5e8933b3b0" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.615255 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e3380ad2b764a851403a900b5b67bc0bfce8ec5b8d3c950aa0dbb5e8933b3b0"} err="failed to get container status \"8e3380ad2b764a851403a900b5b67bc0bfce8ec5b8d3c950aa0dbb5e8933b3b0\": rpc error: code = NotFound desc = could not find container \"8e3380ad2b764a851403a900b5b67bc0bfce8ec5b8d3c950aa0dbb5e8933b3b0\": container with ID starting with 8e3380ad2b764a851403a900b5b67bc0bfce8ec5b8d3c950aa0dbb5e8933b3b0 not found: ID does not exist" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.615297 4744 scope.go:117] "RemoveContainer" containerID="9c3ae761e81730a0da4906c9dc39fb77f17c2ca8027ca2a411bd2340fabc141b" Oct 08 09:15:16 crc kubenswrapper[4744]: E1008 09:15:16.615693 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c3ae761e81730a0da4906c9dc39fb77f17c2ca8027ca2a411bd2340fabc141b\": container with ID starting with 9c3ae761e81730a0da4906c9dc39fb77f17c2ca8027ca2a411bd2340fabc141b not found: ID does not exist" containerID="9c3ae761e81730a0da4906c9dc39fb77f17c2ca8027ca2a411bd2340fabc141b" Oct 08 09:15:16 crc kubenswrapper[4744]: I1008 09:15:16.615723 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c3ae761e81730a0da4906c9dc39fb77f17c2ca8027ca2a411bd2340fabc141b"} err="failed to get container status \"9c3ae761e81730a0da4906c9dc39fb77f17c2ca8027ca2a411bd2340fabc141b\": rpc error: code = NotFound desc = could not find container \"9c3ae761e81730a0da4906c9dc39fb77f17c2ca8027ca2a411bd2340fabc141b\": container with ID starting with 9c3ae761e81730a0da4906c9dc39fb77f17c2ca8027ca2a411bd2340fabc141b not found: ID does not exist" Oct 08 09:15:17 crc kubenswrapper[4744]: I1008 09:15:17.465231 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baee305b-d6aa-41ac-a360-77ceb69f5db0" path="/var/lib/kubelet/pods/baee305b-d6aa-41ac-a360-77ceb69f5db0/volumes" Oct 08 09:15:17 crc kubenswrapper[4744]: I1008 09:15:17.467782 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e33828e9-5ea7-4f7d-ab35-474268801800" path="/var/lib/kubelet/pods/e33828e9-5ea7-4f7d-ab35-474268801800/volumes" Oct 08 09:15:17 crc kubenswrapper[4744]: I1008 09:15:17.493050 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-svt5h"] Oct 08 09:15:17 crc kubenswrapper[4744]: I1008 09:15:17.547534 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-svt5h" podUID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" containerName="registry-server" containerID="cri-o://d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897" gracePeriod=2 Oct 08 09:15:17 crc kubenswrapper[4744]: E1008 09:15:17.805671 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc46e61ce_2d24_48e3_bc3a_e9c3c4fab8bc.slice/crio-conmon-d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897.scope\": RecentStats: unable to find data in memory cache]" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.430157 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.556080 4744 generic.go:334] "Generic (PLEG): container finished" podID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" containerID="d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897" exitCode=0 Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.556128 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svt5h" event={"ID":"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc","Type":"ContainerDied","Data":"d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897"} Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.556162 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-svt5h" event={"ID":"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc","Type":"ContainerDied","Data":"4a1de6582f89bb72196fabb35d97c51199834633aa8244ae1b52eeb382ee6fed"} Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.556180 4744 scope.go:117] "RemoveContainer" containerID="d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.556319 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-svt5h" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.572768 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k2zn\" (UniqueName: \"kubernetes.io/projected/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-kube-api-access-2k2zn\") pod \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\" (UID: \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\") " Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.572818 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-utilities\") pod \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\" (UID: \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\") " Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.572858 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-catalog-content\") pod \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\" (UID: \"c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc\") " Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.574253 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-utilities" (OuterVolumeSpecName: "utilities") pod "c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" (UID: "c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.578892 4744 scope.go:117] "RemoveContainer" containerID="602a06c61efbdf18f91dc32c2e37e819b61123ec4be8b2f10e352104d62799db" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.580342 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-kube-api-access-2k2zn" (OuterVolumeSpecName: "kube-api-access-2k2zn") pod "c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" (UID: "c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc"). InnerVolumeSpecName "kube-api-access-2k2zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.590192 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" (UID: "c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.604336 4744 scope.go:117] "RemoveContainer" containerID="652ba64bbd4dc9cbd79d3b648ba3465acd69932cf55c2a67c9831b7b42a3b248" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.628580 4744 scope.go:117] "RemoveContainer" containerID="d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897" Oct 08 09:15:18 crc kubenswrapper[4744]: E1008 09:15:18.631813 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897\": container with ID starting with d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897 not found: ID does not exist" containerID="d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.631923 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897"} err="failed to get container status \"d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897\": rpc error: code = NotFound desc = could not find container \"d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897\": container with ID starting with d9eea1a004406565b305f682e96b19c6429eba1e29089e191658e3764c0dd897 not found: ID does not exist" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.631978 4744 scope.go:117] "RemoveContainer" containerID="602a06c61efbdf18f91dc32c2e37e819b61123ec4be8b2f10e352104d62799db" Oct 08 09:15:18 crc kubenswrapper[4744]: E1008 09:15:18.639518 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"602a06c61efbdf18f91dc32c2e37e819b61123ec4be8b2f10e352104d62799db\": container with ID starting with 602a06c61efbdf18f91dc32c2e37e819b61123ec4be8b2f10e352104d62799db not found: ID does not exist" containerID="602a06c61efbdf18f91dc32c2e37e819b61123ec4be8b2f10e352104d62799db" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.639578 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"602a06c61efbdf18f91dc32c2e37e819b61123ec4be8b2f10e352104d62799db"} err="failed to get container status \"602a06c61efbdf18f91dc32c2e37e819b61123ec4be8b2f10e352104d62799db\": rpc error: code = NotFound desc = could not find container \"602a06c61efbdf18f91dc32c2e37e819b61123ec4be8b2f10e352104d62799db\": container with ID starting with 602a06c61efbdf18f91dc32c2e37e819b61123ec4be8b2f10e352104d62799db not found: ID does not exist" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.639610 4744 scope.go:117] "RemoveContainer" containerID="652ba64bbd4dc9cbd79d3b648ba3465acd69932cf55c2a67c9831b7b42a3b248" Oct 08 09:15:18 crc kubenswrapper[4744]: E1008 09:15:18.640306 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"652ba64bbd4dc9cbd79d3b648ba3465acd69932cf55c2a67c9831b7b42a3b248\": container with ID starting with 652ba64bbd4dc9cbd79d3b648ba3465acd69932cf55c2a67c9831b7b42a3b248 not found: ID does not exist" containerID="652ba64bbd4dc9cbd79d3b648ba3465acd69932cf55c2a67c9831b7b42a3b248" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.640340 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"652ba64bbd4dc9cbd79d3b648ba3465acd69932cf55c2a67c9831b7b42a3b248"} err="failed to get container status \"652ba64bbd4dc9cbd79d3b648ba3465acd69932cf55c2a67c9831b7b42a3b248\": rpc error: code = NotFound desc = could not find container \"652ba64bbd4dc9cbd79d3b648ba3465acd69932cf55c2a67c9831b7b42a3b248\": container with ID starting with 652ba64bbd4dc9cbd79d3b648ba3465acd69932cf55c2a67c9831b7b42a3b248 not found: ID does not exist" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.674864 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k2zn\" (UniqueName: \"kubernetes.io/projected/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-kube-api-access-2k2zn\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.674924 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.674936 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.891438 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-svt5h"] Oct 08 09:15:18 crc kubenswrapper[4744]: I1008 09:15:18.895789 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-svt5h"] Oct 08 09:15:19 crc kubenswrapper[4744]: I1008 09:15:19.460611 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" path="/var/lib/kubelet/pods/c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc/volumes" Oct 08 09:15:19 crc kubenswrapper[4744]: I1008 09:15:19.689923 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:15:19 crc kubenswrapper[4744]: I1008 09:15:19.690002 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:15:19 crc kubenswrapper[4744]: I1008 09:15:19.690064 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:15:19 crc kubenswrapper[4744]: I1008 09:15:19.690890 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 09:15:19 crc kubenswrapper[4744]: I1008 09:15:19.690946 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0" gracePeriod=600 Oct 08 09:15:20 crc kubenswrapper[4744]: I1008 09:15:20.572263 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0" exitCode=0 Oct 08 09:15:20 crc kubenswrapper[4744]: I1008 09:15:20.572386 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0"} Oct 08 09:15:20 crc kubenswrapper[4744]: I1008 09:15:20.572809 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"63e1dfa02860e1e750f72656194fd55b1911a02f76a9cd7c271f9f20e1dffa0a"} Oct 08 09:15:36 crc kubenswrapper[4744]: I1008 09:15:36.578518 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" podUID="14215523-611c-4c7d-a2a9-f4dbe478b0e0" containerName="oauth-openshift" containerID="cri-o://4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809" gracePeriod=15 Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.018275 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.071829 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7f54ff7574-kl5hf"] Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072145 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e33828e9-5ea7-4f7d-ab35-474268801800" containerName="extract-content" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072164 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e33828e9-5ea7-4f7d-ab35-474268801800" containerName="extract-content" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072182 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e33828e9-5ea7-4f7d-ab35-474268801800" containerName="extract-utilities" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072190 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e33828e9-5ea7-4f7d-ab35-474268801800" containerName="extract-utilities" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072204 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baee305b-d6aa-41ac-a360-77ceb69f5db0" containerName="extract-content" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072213 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="baee305b-d6aa-41ac-a360-77ceb69f5db0" containerName="extract-content" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072223 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" containerName="registry-server" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072232 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" containerName="registry-server" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072243 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc8f9d4-5df0-420c-9cba-aae825a33836" containerName="extract-utilities" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072251 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc8f9d4-5df0-420c-9cba-aae825a33836" containerName="extract-utilities" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072263 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" containerName="extract-utilities" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072270 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" containerName="extract-utilities" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072285 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e33828e9-5ea7-4f7d-ab35-474268801800" containerName="registry-server" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072292 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e33828e9-5ea7-4f7d-ab35-474268801800" containerName="registry-server" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072308 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" containerName="extract-content" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072316 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" containerName="extract-content" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072330 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc8f9d4-5df0-420c-9cba-aae825a33836" containerName="extract-content" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072338 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc8f9d4-5df0-420c-9cba-aae825a33836" containerName="extract-content" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072349 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baee305b-d6aa-41ac-a360-77ceb69f5db0" containerName="extract-utilities" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072357 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="baee305b-d6aa-41ac-a360-77ceb69f5db0" containerName="extract-utilities" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072395 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14215523-611c-4c7d-a2a9-f4dbe478b0e0" containerName="oauth-openshift" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072406 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="14215523-611c-4c7d-a2a9-f4dbe478b0e0" containerName="oauth-openshift" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072415 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c847191-fb16-4cdc-8b54-ee09e3571cce" containerName="collect-profiles" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072422 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c847191-fb16-4cdc-8b54-ee09e3571cce" containerName="collect-profiles" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072433 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baee305b-d6aa-41ac-a360-77ceb69f5db0" containerName="registry-server" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072441 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="baee305b-d6aa-41ac-a360-77ceb69f5db0" containerName="registry-server" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.072452 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc8f9d4-5df0-420c-9cba-aae825a33836" containerName="registry-server" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072460 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc8f9d4-5df0-420c-9cba-aae825a33836" containerName="registry-server" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072783 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c46e61ce-2d24-48e3-bc3a-e9c3c4fab8bc" containerName="registry-server" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072796 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e33828e9-5ea7-4f7d-ab35-474268801800" containerName="registry-server" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072808 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c847191-fb16-4cdc-8b54-ee09e3571cce" containerName="collect-profiles" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072819 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="baee305b-d6aa-41ac-a360-77ceb69f5db0" containerName="registry-server" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072831 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="14215523-611c-4c7d-a2a9-f4dbe478b0e0" containerName="oauth-openshift" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.072843 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc8f9d4-5df0-420c-9cba-aae825a33836" containerName="registry-server" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.073437 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.097889 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7f54ff7574-kl5hf"] Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164192 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-cliconfig\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164252 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-login\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164279 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-audit-policies\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164307 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-ocp-branding-template\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164369 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-trusted-ca-bundle\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164436 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-idp-0-file-data\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164462 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-service-ca\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164480 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-session\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164524 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14215523-611c-4c7d-a2a9-f4dbe478b0e0-audit-dir\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164547 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-serving-cert\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164586 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m76cx\" (UniqueName: \"kubernetes.io/projected/14215523-611c-4c7d-a2a9-f4dbe478b0e0-kube-api-access-m76cx\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164609 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-error\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164643 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-provider-selection\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.164666 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-router-certs\") pod \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\" (UID: \"14215523-611c-4c7d-a2a9-f4dbe478b0e0\") " Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.165120 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.165886 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.165970 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14215523-611c-4c7d-a2a9-f4dbe478b0e0-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.166107 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.166435 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.171158 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14215523-611c-4c7d-a2a9-f4dbe478b0e0-kube-api-access-m76cx" (OuterVolumeSpecName: "kube-api-access-m76cx") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "kube-api-access-m76cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.171202 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.171226 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.171560 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.171822 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.172676 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.173313 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.174634 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.178615 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "14215523-611c-4c7d-a2a9-f4dbe478b0e0" (UID: "14215523-611c-4c7d-a2a9-f4dbe478b0e0"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.266531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-service-ca\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.266580 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.266602 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-user-template-login\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.266625 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-audit-dir\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.266645 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-user-template-error\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.266681 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-audit-policies\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.266706 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-session\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.266849 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267002 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267087 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267141 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267201 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-router-certs\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267244 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmtwq\" (UniqueName: \"kubernetes.io/projected/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-kube-api-access-zmtwq\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267301 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267612 4744 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14215523-611c-4c7d-a2a9-f4dbe478b0e0-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267641 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267664 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m76cx\" (UniqueName: \"kubernetes.io/projected/14215523-611c-4c7d-a2a9-f4dbe478b0e0-kube-api-access-m76cx\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267685 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267710 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267733 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267755 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267783 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267807 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267830 4744 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267853 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267875 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267894 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.267918 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14215523-611c-4c7d-a2a9-f4dbe478b0e0-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369313 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-audit-dir\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369429 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-user-template-error\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369483 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-audit-dir\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369498 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-audit-policies\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369624 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-session\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369648 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369689 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369738 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369761 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369803 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-router-certs\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369822 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmtwq\" (UniqueName: \"kubernetes.io/projected/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-kube-api-access-zmtwq\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369848 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.369983 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-service-ca\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.370010 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.370026 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-user-template-login\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.371214 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-audit-policies\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.372628 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.372985 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.375958 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-user-template-login\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.376576 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.377302 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.377722 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-session\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.377908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-user-template-error\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.378947 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.381148 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-service-ca\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.383077 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-router-certs\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.384755 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.391326 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmtwq\" (UniqueName: \"kubernetes.io/projected/a0ef06b8-675e-4b47-a3f4-f9d6230f83be-kube-api-access-zmtwq\") pod \"oauth-openshift-7f54ff7574-kl5hf\" (UID: \"a0ef06b8-675e-4b47-a3f4-f9d6230f83be\") " pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.397167 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.666355 4744 generic.go:334] "Generic (PLEG): container finished" podID="14215523-611c-4c7d-a2a9-f4dbe478b0e0" containerID="4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809" exitCode=0 Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.666423 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" event={"ID":"14215523-611c-4c7d-a2a9-f4dbe478b0e0","Type":"ContainerDied","Data":"4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809"} Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.666739 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" event={"ID":"14215523-611c-4c7d-a2a9-f4dbe478b0e0","Type":"ContainerDied","Data":"bfe62576cac9deb3b2129069127fbb5a1341becc7b4effc7250ac71dc8341e8f"} Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.666492 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-x97sx" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.666760 4744 scope.go:117] "RemoveContainer" containerID="4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.696795 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x97sx"] Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.700729 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x97sx"] Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.706248 4744 scope.go:117] "RemoveContainer" containerID="4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809" Oct 08 09:15:37 crc kubenswrapper[4744]: E1008 09:15:37.707089 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809\": container with ID starting with 4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809 not found: ID does not exist" containerID="4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.707165 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809"} err="failed to get container status \"4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809\": rpc error: code = NotFound desc = could not find container \"4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809\": container with ID starting with 4c6673702dd92a584756da97c1f338e0803ab60bd656c11b49e71ab73fb06809 not found: ID does not exist" Oct 08 09:15:37 crc kubenswrapper[4744]: I1008 09:15:37.895364 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7f54ff7574-kl5hf"] Oct 08 09:15:38 crc kubenswrapper[4744]: I1008 09:15:38.678894 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" event={"ID":"a0ef06b8-675e-4b47-a3f4-f9d6230f83be","Type":"ContainerStarted","Data":"928efba26a5f902f9858aa8329ff71d06983ba8cd3207c9d90087e9ebf9db095"} Oct 08 09:15:38 crc kubenswrapper[4744]: I1008 09:15:38.679423 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" event={"ID":"a0ef06b8-675e-4b47-a3f4-f9d6230f83be","Type":"ContainerStarted","Data":"68c4570933f50e61fc299ae65b5360123205775499ef07ee23dbbc1d84dd7d18"} Oct 08 09:15:38 crc kubenswrapper[4744]: I1008 09:15:38.681232 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:38 crc kubenswrapper[4744]: I1008 09:15:38.719264 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" podStartSLOduration=27.719223834 podStartE2EDuration="27.719223834s" podCreationTimestamp="2025-10-08 09:15:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:15:38.718778815 +0000 UTC m=+233.966424084" watchObservedRunningTime="2025-10-08 09:15:38.719223834 +0000 UTC m=+233.966869083" Oct 08 09:15:38 crc kubenswrapper[4744]: I1008 09:15:38.749674 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7f54ff7574-kl5hf" Oct 08 09:15:39 crc kubenswrapper[4744]: I1008 09:15:39.463518 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14215523-611c-4c7d-a2a9-f4dbe478b0e0" path="/var/lib/kubelet/pods/14215523-611c-4c7d-a2a9-f4dbe478b0e0/volumes" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.251463 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pvc9m"] Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.252283 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pvc9m" podUID="62e5fd77-7efc-455d-86d0-763bb040f2d0" containerName="registry-server" containerID="cri-o://18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8" gracePeriod=30 Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.261493 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rsqj4"] Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.263185 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rsqj4" podUID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" containerName="registry-server" containerID="cri-o://a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf" gracePeriod=30 Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.277253 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vkldx"] Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.277506 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" podUID="382e1d8e-0916-487b-9f6c-2d946da409f1" containerName="marketplace-operator" containerID="cri-o://397ad7ee1c10dd3e48ac2b8fa0931662bde2f88a10f857d0865d6838def22ad6" gracePeriod=30 Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.289988 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-85kw7"] Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.290342 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-85kw7" podUID="3c40f62a-9591-4187-aaf5-bbf7319d013c" containerName="registry-server" containerID="cri-o://5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9" gracePeriod=30 Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.299505 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6qmwr"] Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.299793 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6qmwr" podUID="432cd400-49d6-4dc1-a40d-182f5c6f4556" containerName="registry-server" containerID="cri-o://043047350a72aeef1dc3d32608103d9ffa7338b8089f926e6e881bd8ae724e14" gracePeriod=30 Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.319648 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rjzwz"] Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.320836 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.345807 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rjzwz"] Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.483493 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0364f287-922c-410f-9cee-1866cdba42e1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rjzwz\" (UID: \"0364f287-922c-410f-9cee-1866cdba42e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.483565 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0364f287-922c-410f-9cee-1866cdba42e1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rjzwz\" (UID: \"0364f287-922c-410f-9cee-1866cdba42e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.483653 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vkj5\" (UniqueName: \"kubernetes.io/projected/0364f287-922c-410f-9cee-1866cdba42e1-kube-api-access-6vkj5\") pod \"marketplace-operator-79b997595-rjzwz\" (UID: \"0364f287-922c-410f-9cee-1866cdba42e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.584842 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vkj5\" (UniqueName: \"kubernetes.io/projected/0364f287-922c-410f-9cee-1866cdba42e1-kube-api-access-6vkj5\") pod \"marketplace-operator-79b997595-rjzwz\" (UID: \"0364f287-922c-410f-9cee-1866cdba42e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.585187 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0364f287-922c-410f-9cee-1866cdba42e1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rjzwz\" (UID: \"0364f287-922c-410f-9cee-1866cdba42e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.585254 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0364f287-922c-410f-9cee-1866cdba42e1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rjzwz\" (UID: \"0364f287-922c-410f-9cee-1866cdba42e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.587150 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0364f287-922c-410f-9cee-1866cdba42e1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rjzwz\" (UID: \"0364f287-922c-410f-9cee-1866cdba42e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.598172 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0364f287-922c-410f-9cee-1866cdba42e1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rjzwz\" (UID: \"0364f287-922c-410f-9cee-1866cdba42e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.603115 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vkj5\" (UniqueName: \"kubernetes.io/projected/0364f287-922c-410f-9cee-1866cdba42e1-kube-api-access-6vkj5\") pod \"marketplace-operator-79b997595-rjzwz\" (UID: \"0364f287-922c-410f-9cee-1866cdba42e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.636485 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.679571 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.746094 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.761474 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.789883 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62e5fd77-7efc-455d-86d0-763bb040f2d0-utilities\") pod \"62e5fd77-7efc-455d-86d0-763bb040f2d0\" (UID: \"62e5fd77-7efc-455d-86d0-763bb040f2d0\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.789940 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62e5fd77-7efc-455d-86d0-763bb040f2d0-catalog-content\") pod \"62e5fd77-7efc-455d-86d0-763bb040f2d0\" (UID: \"62e5fd77-7efc-455d-86d0-763bb040f2d0\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.790016 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5jx6\" (UniqueName: \"kubernetes.io/projected/62e5fd77-7efc-455d-86d0-763bb040f2d0-kube-api-access-w5jx6\") pod \"62e5fd77-7efc-455d-86d0-763bb040f2d0\" (UID: \"62e5fd77-7efc-455d-86d0-763bb040f2d0\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.791303 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62e5fd77-7efc-455d-86d0-763bb040f2d0-utilities" (OuterVolumeSpecName: "utilities") pod "62e5fd77-7efc-455d-86d0-763bb040f2d0" (UID: "62e5fd77-7efc-455d-86d0-763bb040f2d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.792230 4744 generic.go:334] "Generic (PLEG): container finished" podID="3c40f62a-9591-4187-aaf5-bbf7319d013c" containerID="5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9" exitCode=0 Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.792351 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-85kw7" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.792445 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85kw7" event={"ID":"3c40f62a-9591-4187-aaf5-bbf7319d013c","Type":"ContainerDied","Data":"5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9"} Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.792512 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-85kw7" event={"ID":"3c40f62a-9591-4187-aaf5-bbf7319d013c","Type":"ContainerDied","Data":"278dc9a7571c5088961e5d832cc20ea9e23f61954744d30f59fc33e5c7b13330"} Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.792531 4744 scope.go:117] "RemoveContainer" containerID="5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.801583 4744 generic.go:334] "Generic (PLEG): container finished" podID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" containerID="a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf" exitCode=0 Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.801725 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsqj4" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.802423 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsqj4" event={"ID":"d28208cc-1278-45d6-9d2e-a65a6d0c94d8","Type":"ContainerDied","Data":"a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf"} Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.802457 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsqj4" event={"ID":"d28208cc-1278-45d6-9d2e-a65a6d0c94d8","Type":"ContainerDied","Data":"4a8c42d8769c173dce537f986a3900fdeba1609735b0f9b7f5793dd37e7ea71b"} Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.802439 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62e5fd77-7efc-455d-86d0-763bb040f2d0-kube-api-access-w5jx6" (OuterVolumeSpecName: "kube-api-access-w5jx6") pod "62e5fd77-7efc-455d-86d0-763bb040f2d0" (UID: "62e5fd77-7efc-455d-86d0-763bb040f2d0"). InnerVolumeSpecName "kube-api-access-w5jx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.806238 4744 generic.go:334] "Generic (PLEG): container finished" podID="62e5fd77-7efc-455d-86d0-763bb040f2d0" containerID="18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8" exitCode=0 Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.806319 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvc9m" event={"ID":"62e5fd77-7efc-455d-86d0-763bb040f2d0","Type":"ContainerDied","Data":"18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8"} Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.806352 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pvc9m" event={"ID":"62e5fd77-7efc-455d-86d0-763bb040f2d0","Type":"ContainerDied","Data":"5b6a45032810b901a099856881223e0d6809f74970ef7baa6ff6ae145ec84c10"} Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.806329 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pvc9m" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.809343 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.810216 4744 generic.go:334] "Generic (PLEG): container finished" podID="432cd400-49d6-4dc1-a40d-182f5c6f4556" containerID="043047350a72aeef1dc3d32608103d9ffa7338b8089f926e6e881bd8ae724e14" exitCode=0 Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.810263 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qmwr" event={"ID":"432cd400-49d6-4dc1-a40d-182f5c6f4556","Type":"ContainerDied","Data":"043047350a72aeef1dc3d32608103d9ffa7338b8089f926e6e881bd8ae724e14"} Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.813782 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.813942 4744 generic.go:334] "Generic (PLEG): container finished" podID="382e1d8e-0916-487b-9f6c-2d946da409f1" containerID="397ad7ee1c10dd3e48ac2b8fa0931662bde2f88a10f857d0865d6838def22ad6" exitCode=0 Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.813982 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" event={"ID":"382e1d8e-0916-487b-9f6c-2d946da409f1","Type":"ContainerDied","Data":"397ad7ee1c10dd3e48ac2b8fa0931662bde2f88a10f857d0865d6838def22ad6"} Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.857437 4744 scope.go:117] "RemoveContainer" containerID="7b2b5d94668a12b3883c14ad81f8afade08a21e0e39a403309ca7850d2435f49" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.866687 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62e5fd77-7efc-455d-86d0-763bb040f2d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62e5fd77-7efc-455d-86d0-763bb040f2d0" (UID: "62e5fd77-7efc-455d-86d0-763bb040f2d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.891920 4744 scope.go:117] "RemoveContainer" containerID="28bc326074c71f958725ed207f7f6064137213c79eae417113166dc77f9e7d45" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.892166 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr262\" (UniqueName: \"kubernetes.io/projected/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-kube-api-access-zr262\") pod \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\" (UID: \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.892239 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c40f62a-9591-4187-aaf5-bbf7319d013c-utilities\") pod \"3c40f62a-9591-4187-aaf5-bbf7319d013c\" (UID: \"3c40f62a-9591-4187-aaf5-bbf7319d013c\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.892269 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c40f62a-9591-4187-aaf5-bbf7319d013c-catalog-content\") pod \"3c40f62a-9591-4187-aaf5-bbf7319d013c\" (UID: \"3c40f62a-9591-4187-aaf5-bbf7319d013c\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.892300 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c692\" (UniqueName: \"kubernetes.io/projected/3c40f62a-9591-4187-aaf5-bbf7319d013c-kube-api-access-5c692\") pod \"3c40f62a-9591-4187-aaf5-bbf7319d013c\" (UID: \"3c40f62a-9591-4187-aaf5-bbf7319d013c\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.892340 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-utilities\") pod \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\" (UID: \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.892409 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-catalog-content\") pod \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\" (UID: \"d28208cc-1278-45d6-9d2e-a65a6d0c94d8\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.892592 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62e5fd77-7efc-455d-86d0-763bb040f2d0-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.892602 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62e5fd77-7efc-455d-86d0-763bb040f2d0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.892613 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5jx6\" (UniqueName: \"kubernetes.io/projected/62e5fd77-7efc-455d-86d0-763bb040f2d0-kube-api-access-w5jx6\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.893356 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c40f62a-9591-4187-aaf5-bbf7319d013c-utilities" (OuterVolumeSpecName: "utilities") pod "3c40f62a-9591-4187-aaf5-bbf7319d013c" (UID: "3c40f62a-9591-4187-aaf5-bbf7319d013c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.896057 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-utilities" (OuterVolumeSpecName: "utilities") pod "d28208cc-1278-45d6-9d2e-a65a6d0c94d8" (UID: "d28208cc-1278-45d6-9d2e-a65a6d0c94d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.896316 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c40f62a-9591-4187-aaf5-bbf7319d013c-kube-api-access-5c692" (OuterVolumeSpecName: "kube-api-access-5c692") pod "3c40f62a-9591-4187-aaf5-bbf7319d013c" (UID: "3c40f62a-9591-4187-aaf5-bbf7319d013c"). InnerVolumeSpecName "kube-api-access-5c692". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.897670 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-kube-api-access-zr262" (OuterVolumeSpecName: "kube-api-access-zr262") pod "d28208cc-1278-45d6-9d2e-a65a6d0c94d8" (UID: "d28208cc-1278-45d6-9d2e-a65a6d0c94d8"). InnerVolumeSpecName "kube-api-access-zr262". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.932936 4744 scope.go:117] "RemoveContainer" containerID="5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9" Oct 08 09:15:51 crc kubenswrapper[4744]: E1008 09:15:51.934116 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9\": container with ID starting with 5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9 not found: ID does not exist" containerID="5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.934243 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9"} err="failed to get container status \"5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9\": rpc error: code = NotFound desc = could not find container \"5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9\": container with ID starting with 5ce09624a1999e4d56af12917f6cbaf7363610c8a49db6e3d90fb595251b7ed9 not found: ID does not exist" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.934357 4744 scope.go:117] "RemoveContainer" containerID="7b2b5d94668a12b3883c14ad81f8afade08a21e0e39a403309ca7850d2435f49" Oct 08 09:15:51 crc kubenswrapper[4744]: E1008 09:15:51.934993 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b2b5d94668a12b3883c14ad81f8afade08a21e0e39a403309ca7850d2435f49\": container with ID starting with 7b2b5d94668a12b3883c14ad81f8afade08a21e0e39a403309ca7850d2435f49 not found: ID does not exist" containerID="7b2b5d94668a12b3883c14ad81f8afade08a21e0e39a403309ca7850d2435f49" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.935092 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b2b5d94668a12b3883c14ad81f8afade08a21e0e39a403309ca7850d2435f49"} err="failed to get container status \"7b2b5d94668a12b3883c14ad81f8afade08a21e0e39a403309ca7850d2435f49\": rpc error: code = NotFound desc = could not find container \"7b2b5d94668a12b3883c14ad81f8afade08a21e0e39a403309ca7850d2435f49\": container with ID starting with 7b2b5d94668a12b3883c14ad81f8afade08a21e0e39a403309ca7850d2435f49 not found: ID does not exist" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.935241 4744 scope.go:117] "RemoveContainer" containerID="28bc326074c71f958725ed207f7f6064137213c79eae417113166dc77f9e7d45" Oct 08 09:15:51 crc kubenswrapper[4744]: E1008 09:15:51.937035 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28bc326074c71f958725ed207f7f6064137213c79eae417113166dc77f9e7d45\": container with ID starting with 28bc326074c71f958725ed207f7f6064137213c79eae417113166dc77f9e7d45 not found: ID does not exist" containerID="28bc326074c71f958725ed207f7f6064137213c79eae417113166dc77f9e7d45" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.937128 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28bc326074c71f958725ed207f7f6064137213c79eae417113166dc77f9e7d45"} err="failed to get container status \"28bc326074c71f958725ed207f7f6064137213c79eae417113166dc77f9e7d45\": rpc error: code = NotFound desc = could not find container \"28bc326074c71f958725ed207f7f6064137213c79eae417113166dc77f9e7d45\": container with ID starting with 28bc326074c71f958725ed207f7f6064137213c79eae417113166dc77f9e7d45 not found: ID does not exist" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.937189 4744 scope.go:117] "RemoveContainer" containerID="a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.937645 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c40f62a-9591-4187-aaf5-bbf7319d013c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c40f62a-9591-4187-aaf5-bbf7319d013c" (UID: "3c40f62a-9591-4187-aaf5-bbf7319d013c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.949404 4744 scope.go:117] "RemoveContainer" containerID="d1304c2c5cf4962ce5d69fd4e8492821e33636b7574545b48c877e49e8850060" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.963594 4744 scope.go:117] "RemoveContainer" containerID="152037bb0738eb0e0dfdf4d4d6199a6e2bc250a722a15ceadaf0e0d81fd712db" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.968970 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d28208cc-1278-45d6-9d2e-a65a6d0c94d8" (UID: "d28208cc-1278-45d6-9d2e-a65a6d0c94d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.979841 4744 scope.go:117] "RemoveContainer" containerID="a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf" Oct 08 09:15:51 crc kubenswrapper[4744]: E1008 09:15:51.980592 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf\": container with ID starting with a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf not found: ID does not exist" containerID="a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.980647 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf"} err="failed to get container status \"a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf\": rpc error: code = NotFound desc = could not find container \"a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf\": container with ID starting with a8cd2bbf255bde5f88f5f53a781ff98812767374cb59a7c5d713f433f2a87caf not found: ID does not exist" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.980699 4744 scope.go:117] "RemoveContainer" containerID="d1304c2c5cf4962ce5d69fd4e8492821e33636b7574545b48c877e49e8850060" Oct 08 09:15:51 crc kubenswrapper[4744]: E1008 09:15:51.981133 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1304c2c5cf4962ce5d69fd4e8492821e33636b7574545b48c877e49e8850060\": container with ID starting with d1304c2c5cf4962ce5d69fd4e8492821e33636b7574545b48c877e49e8850060 not found: ID does not exist" containerID="d1304c2c5cf4962ce5d69fd4e8492821e33636b7574545b48c877e49e8850060" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.981232 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1304c2c5cf4962ce5d69fd4e8492821e33636b7574545b48c877e49e8850060"} err="failed to get container status \"d1304c2c5cf4962ce5d69fd4e8492821e33636b7574545b48c877e49e8850060\": rpc error: code = NotFound desc = could not find container \"d1304c2c5cf4962ce5d69fd4e8492821e33636b7574545b48c877e49e8850060\": container with ID starting with d1304c2c5cf4962ce5d69fd4e8492821e33636b7574545b48c877e49e8850060 not found: ID does not exist" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.981327 4744 scope.go:117] "RemoveContainer" containerID="152037bb0738eb0e0dfdf4d4d6199a6e2bc250a722a15ceadaf0e0d81fd712db" Oct 08 09:15:51 crc kubenswrapper[4744]: E1008 09:15:51.981641 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"152037bb0738eb0e0dfdf4d4d6199a6e2bc250a722a15ceadaf0e0d81fd712db\": container with ID starting with 152037bb0738eb0e0dfdf4d4d6199a6e2bc250a722a15ceadaf0e0d81fd712db not found: ID does not exist" containerID="152037bb0738eb0e0dfdf4d4d6199a6e2bc250a722a15ceadaf0e0d81fd712db" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.981667 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"152037bb0738eb0e0dfdf4d4d6199a6e2bc250a722a15ceadaf0e0d81fd712db"} err="failed to get container status \"152037bb0738eb0e0dfdf4d4d6199a6e2bc250a722a15ceadaf0e0d81fd712db\": rpc error: code = NotFound desc = could not find container \"152037bb0738eb0e0dfdf4d4d6199a6e2bc250a722a15ceadaf0e0d81fd712db\": container with ID starting with 152037bb0738eb0e0dfdf4d4d6199a6e2bc250a722a15ceadaf0e0d81fd712db not found: ID does not exist" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.981692 4744 scope.go:117] "RemoveContainer" containerID="18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.993542 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/432cd400-49d6-4dc1-a40d-182f5c6f4556-utilities\") pod \"432cd400-49d6-4dc1-a40d-182f5c6f4556\" (UID: \"432cd400-49d6-4dc1-a40d-182f5c6f4556\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.993672 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4kxf\" (UniqueName: \"kubernetes.io/projected/382e1d8e-0916-487b-9f6c-2d946da409f1-kube-api-access-k4kxf\") pod \"382e1d8e-0916-487b-9f6c-2d946da409f1\" (UID: \"382e1d8e-0916-487b-9f6c-2d946da409f1\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.993842 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/432cd400-49d6-4dc1-a40d-182f5c6f4556-catalog-content\") pod \"432cd400-49d6-4dc1-a40d-182f5c6f4556\" (UID: \"432cd400-49d6-4dc1-a40d-182f5c6f4556\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.998462 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/382e1d8e-0916-487b-9f6c-2d946da409f1-marketplace-trusted-ca\") pod \"382e1d8e-0916-487b-9f6c-2d946da409f1\" (UID: \"382e1d8e-0916-487b-9f6c-2d946da409f1\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.998501 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6jjv\" (UniqueName: \"kubernetes.io/projected/432cd400-49d6-4dc1-a40d-182f5c6f4556-kube-api-access-s6jjv\") pod \"432cd400-49d6-4dc1-a40d-182f5c6f4556\" (UID: \"432cd400-49d6-4dc1-a40d-182f5c6f4556\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.998549 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/382e1d8e-0916-487b-9f6c-2d946da409f1-marketplace-operator-metrics\") pod \"382e1d8e-0916-487b-9f6c-2d946da409f1\" (UID: \"382e1d8e-0916-487b-9f6c-2d946da409f1\") " Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.998782 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c40f62a-9591-4187-aaf5-bbf7319d013c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.998796 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c692\" (UniqueName: \"kubernetes.io/projected/3c40f62a-9591-4187-aaf5-bbf7319d013c-kube-api-access-5c692\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.998807 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.998815 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.998823 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr262\" (UniqueName: \"kubernetes.io/projected/d28208cc-1278-45d6-9d2e-a65a6d0c94d8-kube-api-access-zr262\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.998831 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c40f62a-9591-4187-aaf5-bbf7319d013c-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.996113 4744 scope.go:117] "RemoveContainer" containerID="a07b9b00942b4a4f84cb103c3bb84780bf39d126b98e0f6e3a803c9cca73e381" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.994216 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/432cd400-49d6-4dc1-a40d-182f5c6f4556-utilities" (OuterVolumeSpecName: "utilities") pod "432cd400-49d6-4dc1-a40d-182f5c6f4556" (UID: "432cd400-49d6-4dc1-a40d-182f5c6f4556"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.996378 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/382e1d8e-0916-487b-9f6c-2d946da409f1-kube-api-access-k4kxf" (OuterVolumeSpecName: "kube-api-access-k4kxf") pod "382e1d8e-0916-487b-9f6c-2d946da409f1" (UID: "382e1d8e-0916-487b-9f6c-2d946da409f1"). InnerVolumeSpecName "kube-api-access-k4kxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:15:51 crc kubenswrapper[4744]: I1008 09:15:51.999243 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/382e1d8e-0916-487b-9f6c-2d946da409f1-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "382e1d8e-0916-487b-9f6c-2d946da409f1" (UID: "382e1d8e-0916-487b-9f6c-2d946da409f1"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.000859 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/382e1d8e-0916-487b-9f6c-2d946da409f1-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "382e1d8e-0916-487b-9f6c-2d946da409f1" (UID: "382e1d8e-0916-487b-9f6c-2d946da409f1"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.003959 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/432cd400-49d6-4dc1-a40d-182f5c6f4556-kube-api-access-s6jjv" (OuterVolumeSpecName: "kube-api-access-s6jjv") pod "432cd400-49d6-4dc1-a40d-182f5c6f4556" (UID: "432cd400-49d6-4dc1-a40d-182f5c6f4556"). InnerVolumeSpecName "kube-api-access-s6jjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.017622 4744 scope.go:117] "RemoveContainer" containerID="400cc1edc37956ecdf02470cf0483654bd4f02225ee714ec8e71e330d08de4be" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.029279 4744 scope.go:117] "RemoveContainer" containerID="18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8" Oct 08 09:15:52 crc kubenswrapper[4744]: E1008 09:15:52.029689 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8\": container with ID starting with 18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8 not found: ID does not exist" containerID="18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.029716 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8"} err="failed to get container status \"18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8\": rpc error: code = NotFound desc = could not find container \"18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8\": container with ID starting with 18eb07d055663de8743d0e5ee55bbe94e8c0fa93be2b93ba35a06814c2d388b8 not found: ID does not exist" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.029743 4744 scope.go:117] "RemoveContainer" containerID="a07b9b00942b4a4f84cb103c3bb84780bf39d126b98e0f6e3a803c9cca73e381" Oct 08 09:15:52 crc kubenswrapper[4744]: E1008 09:15:52.029958 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a07b9b00942b4a4f84cb103c3bb84780bf39d126b98e0f6e3a803c9cca73e381\": container with ID starting with a07b9b00942b4a4f84cb103c3bb84780bf39d126b98e0f6e3a803c9cca73e381 not found: ID does not exist" containerID="a07b9b00942b4a4f84cb103c3bb84780bf39d126b98e0f6e3a803c9cca73e381" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.029982 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a07b9b00942b4a4f84cb103c3bb84780bf39d126b98e0f6e3a803c9cca73e381"} err="failed to get container status \"a07b9b00942b4a4f84cb103c3bb84780bf39d126b98e0f6e3a803c9cca73e381\": rpc error: code = NotFound desc = could not find container \"a07b9b00942b4a4f84cb103c3bb84780bf39d126b98e0f6e3a803c9cca73e381\": container with ID starting with a07b9b00942b4a4f84cb103c3bb84780bf39d126b98e0f6e3a803c9cca73e381 not found: ID does not exist" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.030002 4744 scope.go:117] "RemoveContainer" containerID="400cc1edc37956ecdf02470cf0483654bd4f02225ee714ec8e71e330d08de4be" Oct 08 09:15:52 crc kubenswrapper[4744]: E1008 09:15:52.030231 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"400cc1edc37956ecdf02470cf0483654bd4f02225ee714ec8e71e330d08de4be\": container with ID starting with 400cc1edc37956ecdf02470cf0483654bd4f02225ee714ec8e71e330d08de4be not found: ID does not exist" containerID="400cc1edc37956ecdf02470cf0483654bd4f02225ee714ec8e71e330d08de4be" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.030251 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"400cc1edc37956ecdf02470cf0483654bd4f02225ee714ec8e71e330d08de4be"} err="failed to get container status \"400cc1edc37956ecdf02470cf0483654bd4f02225ee714ec8e71e330d08de4be\": rpc error: code = NotFound desc = could not find container \"400cc1edc37956ecdf02470cf0483654bd4f02225ee714ec8e71e330d08de4be\": container with ID starting with 400cc1edc37956ecdf02470cf0483654bd4f02225ee714ec8e71e330d08de4be not found: ID does not exist" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.030265 4744 scope.go:117] "RemoveContainer" containerID="043047350a72aeef1dc3d32608103d9ffa7338b8089f926e6e881bd8ae724e14" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.043064 4744 scope.go:117] "RemoveContainer" containerID="0488b56ca4ce620970a8f3b16fef6a6a7dcc90dd757bed319d90e6f938e08cb4" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.056511 4744 scope.go:117] "RemoveContainer" containerID="35477dd9004849232898b38af24d8de0cf816909cbb232e7233bc617a12c7253" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.069770 4744 scope.go:117] "RemoveContainer" containerID="397ad7ee1c10dd3e48ac2b8fa0931662bde2f88a10f857d0865d6838def22ad6" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.075388 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/432cd400-49d6-4dc1-a40d-182f5c6f4556-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "432cd400-49d6-4dc1-a40d-182f5c6f4556" (UID: "432cd400-49d6-4dc1-a40d-182f5c6f4556"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.099492 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/432cd400-49d6-4dc1-a40d-182f5c6f4556-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.099530 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4kxf\" (UniqueName: \"kubernetes.io/projected/382e1d8e-0916-487b-9f6c-2d946da409f1-kube-api-access-k4kxf\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.099540 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/432cd400-49d6-4dc1-a40d-182f5c6f4556-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.099552 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/382e1d8e-0916-487b-9f6c-2d946da409f1-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.099560 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6jjv\" (UniqueName: \"kubernetes.io/projected/432cd400-49d6-4dc1-a40d-182f5c6f4556-kube-api-access-s6jjv\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.099569 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/382e1d8e-0916-487b-9f6c-2d946da409f1-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.120624 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-85kw7"] Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.132770 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-85kw7"] Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.141517 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rsqj4"] Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.148124 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rsqj4"] Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.149147 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rjzwz"] Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.151940 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pvc9m"] Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.154451 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pvc9m"] Oct 08 09:15:52 crc kubenswrapper[4744]: W1008 09:15:52.155381 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0364f287_922c_410f_9cee_1866cdba42e1.slice/crio-e650483448945a3699c55767a515c1df77c3848b5a377ad2cb4e37bc73f43fd2 WatchSource:0}: Error finding container e650483448945a3699c55767a515c1df77c3848b5a377ad2cb4e37bc73f43fd2: Status 404 returned error can't find the container with id e650483448945a3699c55767a515c1df77c3848b5a377ad2cb4e37bc73f43fd2 Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.821567 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" event={"ID":"0364f287-922c-410f-9cee-1866cdba42e1","Type":"ContainerStarted","Data":"ca65e3816ec9a7fa3f0046d8b69b9234a4448c69e6d2c3564a00c23f20a6511c"} Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.822087 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" event={"ID":"0364f287-922c-410f-9cee-1866cdba42e1","Type":"ContainerStarted","Data":"e650483448945a3699c55767a515c1df77c3848b5a377ad2cb4e37bc73f43fd2"} Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.822224 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.822584 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6qmwr" event={"ID":"432cd400-49d6-4dc1-a40d-182f5c6f4556","Type":"ContainerDied","Data":"c39d76e55f6712baa8d739727d1cc0f743e3663d71c16730573e287b2fd93a9c"} Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.822701 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6qmwr" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.824524 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" event={"ID":"382e1d8e-0916-487b-9f6c-2d946da409f1","Type":"ContainerDied","Data":"93ac3e459004cd6a6c2a7ef990882a71ee0bea13068b397dea6e467747f062fb"} Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.824646 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vkldx" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.825726 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.864662 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rjzwz" podStartSLOduration=1.8646426059999999 podStartE2EDuration="1.864642606s" podCreationTimestamp="2025-10-08 09:15:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:15:52.843148303 +0000 UTC m=+248.090793552" watchObservedRunningTime="2025-10-08 09:15:52.864642606 +0000 UTC m=+248.112287845" Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.877859 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6qmwr"] Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.886572 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6qmwr"] Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.909644 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vkldx"] Oct 08 09:15:52 crc kubenswrapper[4744]: I1008 09:15:52.912973 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vkldx"] Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.459407 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="382e1d8e-0916-487b-9f6c-2d946da409f1" path="/var/lib/kubelet/pods/382e1d8e-0916-487b-9f6c-2d946da409f1/volumes" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.459904 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c40f62a-9591-4187-aaf5-bbf7319d013c" path="/var/lib/kubelet/pods/3c40f62a-9591-4187-aaf5-bbf7319d013c/volumes" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.460508 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="432cd400-49d6-4dc1-a40d-182f5c6f4556" path="/var/lib/kubelet/pods/432cd400-49d6-4dc1-a40d-182f5c6f4556/volumes" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.461534 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62e5fd77-7efc-455d-86d0-763bb040f2d0" path="/var/lib/kubelet/pods/62e5fd77-7efc-455d-86d0-763bb040f2d0/volumes" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.462132 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" path="/var/lib/kubelet/pods/d28208cc-1278-45d6-9d2e-a65a6d0c94d8/volumes" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.474734 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g8s9r"] Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.474975 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c40f62a-9591-4187-aaf5-bbf7319d013c" containerName="extract-utilities" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.474991 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c40f62a-9591-4187-aaf5-bbf7319d013c" containerName="extract-utilities" Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.474999 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" containerName="registry-server" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475006 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" containerName="registry-server" Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.475019 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e5fd77-7efc-455d-86d0-763bb040f2d0" containerName="extract-utilities" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475025 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e5fd77-7efc-455d-86d0-763bb040f2d0" containerName="extract-utilities" Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.475036 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c40f62a-9591-4187-aaf5-bbf7319d013c" containerName="registry-server" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475042 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c40f62a-9591-4187-aaf5-bbf7319d013c" containerName="registry-server" Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.475050 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="382e1d8e-0916-487b-9f6c-2d946da409f1" containerName="marketplace-operator" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475056 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="382e1d8e-0916-487b-9f6c-2d946da409f1" containerName="marketplace-operator" Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.475062 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e5fd77-7efc-455d-86d0-763bb040f2d0" containerName="extract-content" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475068 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e5fd77-7efc-455d-86d0-763bb040f2d0" containerName="extract-content" Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.475075 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" containerName="extract-utilities" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475080 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" containerName="extract-utilities" Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.475088 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="432cd400-49d6-4dc1-a40d-182f5c6f4556" containerName="extract-utilities" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475093 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="432cd400-49d6-4dc1-a40d-182f5c6f4556" containerName="extract-utilities" Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.475101 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c40f62a-9591-4187-aaf5-bbf7319d013c" containerName="extract-content" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475108 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c40f62a-9591-4187-aaf5-bbf7319d013c" containerName="extract-content" Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.475114 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62e5fd77-7efc-455d-86d0-763bb040f2d0" containerName="registry-server" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475120 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="62e5fd77-7efc-455d-86d0-763bb040f2d0" containerName="registry-server" Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.475128 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="432cd400-49d6-4dc1-a40d-182f5c6f4556" containerName="registry-server" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475134 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="432cd400-49d6-4dc1-a40d-182f5c6f4556" containerName="registry-server" Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.475142 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="432cd400-49d6-4dc1-a40d-182f5c6f4556" containerName="extract-content" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475147 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="432cd400-49d6-4dc1-a40d-182f5c6f4556" containerName="extract-content" Oct 08 09:15:53 crc kubenswrapper[4744]: E1008 09:15:53.475154 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" containerName="extract-content" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475159 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" containerName="extract-content" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475242 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="382e1d8e-0916-487b-9f6c-2d946da409f1" containerName="marketplace-operator" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475275 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c40f62a-9591-4187-aaf5-bbf7319d013c" containerName="registry-server" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475286 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d28208cc-1278-45d6-9d2e-a65a6d0c94d8" containerName="registry-server" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475299 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="62e5fd77-7efc-455d-86d0-763bb040f2d0" containerName="registry-server" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.475309 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="432cd400-49d6-4dc1-a40d-182f5c6f4556" containerName="registry-server" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.476394 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.480253 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.486467 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g8s9r"] Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.617714 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8a122a-bcbf-474b-8c68-ed54aec6a6f4-utilities\") pod \"redhat-marketplace-g8s9r\" (UID: \"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4\") " pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.617980 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8a122a-bcbf-474b-8c68-ed54aec6a6f4-catalog-content\") pod \"redhat-marketplace-g8s9r\" (UID: \"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4\") " pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.618031 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcl8r\" (UniqueName: \"kubernetes.io/projected/8c8a122a-bcbf-474b-8c68-ed54aec6a6f4-kube-api-access-wcl8r\") pod \"redhat-marketplace-g8s9r\" (UID: \"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4\") " pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.673965 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5gqrp"] Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.675972 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.678723 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.686034 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5gqrp"] Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.719065 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8a122a-bcbf-474b-8c68-ed54aec6a6f4-utilities\") pod \"redhat-marketplace-g8s9r\" (UID: \"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4\") " pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.719122 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8a122a-bcbf-474b-8c68-ed54aec6a6f4-catalog-content\") pod \"redhat-marketplace-g8s9r\" (UID: \"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4\") " pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.719186 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcl8r\" (UniqueName: \"kubernetes.io/projected/8c8a122a-bcbf-474b-8c68-ed54aec6a6f4-kube-api-access-wcl8r\") pod \"redhat-marketplace-g8s9r\" (UID: \"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4\") " pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.720025 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8a122a-bcbf-474b-8c68-ed54aec6a6f4-utilities\") pod \"redhat-marketplace-g8s9r\" (UID: \"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4\") " pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.720299 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8a122a-bcbf-474b-8c68-ed54aec6a6f4-catalog-content\") pod \"redhat-marketplace-g8s9r\" (UID: \"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4\") " pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.742740 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcl8r\" (UniqueName: \"kubernetes.io/projected/8c8a122a-bcbf-474b-8c68-ed54aec6a6f4-kube-api-access-wcl8r\") pod \"redhat-marketplace-g8s9r\" (UID: \"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4\") " pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.807631 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.824249 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtk5c\" (UniqueName: \"kubernetes.io/projected/003a8292-d26b-47f8-89ce-60165fcad99e-kube-api-access-dtk5c\") pod \"certified-operators-5gqrp\" (UID: \"003a8292-d26b-47f8-89ce-60165fcad99e\") " pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.824307 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/003a8292-d26b-47f8-89ce-60165fcad99e-utilities\") pod \"certified-operators-5gqrp\" (UID: \"003a8292-d26b-47f8-89ce-60165fcad99e\") " pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.824392 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/003a8292-d26b-47f8-89ce-60165fcad99e-catalog-content\") pod \"certified-operators-5gqrp\" (UID: \"003a8292-d26b-47f8-89ce-60165fcad99e\") " pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.925183 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/003a8292-d26b-47f8-89ce-60165fcad99e-catalog-content\") pod \"certified-operators-5gqrp\" (UID: \"003a8292-d26b-47f8-89ce-60165fcad99e\") " pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.925285 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtk5c\" (UniqueName: \"kubernetes.io/projected/003a8292-d26b-47f8-89ce-60165fcad99e-kube-api-access-dtk5c\") pod \"certified-operators-5gqrp\" (UID: \"003a8292-d26b-47f8-89ce-60165fcad99e\") " pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.925330 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/003a8292-d26b-47f8-89ce-60165fcad99e-utilities\") pod \"certified-operators-5gqrp\" (UID: \"003a8292-d26b-47f8-89ce-60165fcad99e\") " pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.926293 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/003a8292-d26b-47f8-89ce-60165fcad99e-catalog-content\") pod \"certified-operators-5gqrp\" (UID: \"003a8292-d26b-47f8-89ce-60165fcad99e\") " pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.926504 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/003a8292-d26b-47f8-89ce-60165fcad99e-utilities\") pod \"certified-operators-5gqrp\" (UID: \"003a8292-d26b-47f8-89ce-60165fcad99e\") " pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:15:53 crc kubenswrapper[4744]: I1008 09:15:53.948198 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtk5c\" (UniqueName: \"kubernetes.io/projected/003a8292-d26b-47f8-89ce-60165fcad99e-kube-api-access-dtk5c\") pod \"certified-operators-5gqrp\" (UID: \"003a8292-d26b-47f8-89ce-60165fcad99e\") " pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:15:54 crc kubenswrapper[4744]: I1008 09:15:54.001647 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:15:54 crc kubenswrapper[4744]: I1008 09:15:54.029233 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g8s9r"] Oct 08 09:15:54 crc kubenswrapper[4744]: W1008 09:15:54.035358 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c8a122a_bcbf_474b_8c68_ed54aec6a6f4.slice/crio-db09d66d8bd75a67abdddc962a3dcf52ba2189b1c8168f70187c8e4e66e82e0c WatchSource:0}: Error finding container db09d66d8bd75a67abdddc962a3dcf52ba2189b1c8168f70187c8e4e66e82e0c: Status 404 returned error can't find the container with id db09d66d8bd75a67abdddc962a3dcf52ba2189b1c8168f70187c8e4e66e82e0c Oct 08 09:15:54 crc kubenswrapper[4744]: I1008 09:15:54.217850 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5gqrp"] Oct 08 09:15:54 crc kubenswrapper[4744]: W1008 09:15:54.225128 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod003a8292_d26b_47f8_89ce_60165fcad99e.slice/crio-49fda692724e8db94088ec998930053c1646ec68f6d5c89f7caf6473ba304993 WatchSource:0}: Error finding container 49fda692724e8db94088ec998930053c1646ec68f6d5c89f7caf6473ba304993: Status 404 returned error can't find the container with id 49fda692724e8db94088ec998930053c1646ec68f6d5c89f7caf6473ba304993 Oct 08 09:15:54 crc kubenswrapper[4744]: I1008 09:15:54.857640 4744 generic.go:334] "Generic (PLEG): container finished" podID="8c8a122a-bcbf-474b-8c68-ed54aec6a6f4" containerID="8d4ff64d46c0732794e1f82a6bead5e36076e407b5e2fbae1df984f6b4abb86b" exitCode=0 Oct 08 09:15:54 crc kubenswrapper[4744]: I1008 09:15:54.857691 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8s9r" event={"ID":"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4","Type":"ContainerDied","Data":"8d4ff64d46c0732794e1f82a6bead5e36076e407b5e2fbae1df984f6b4abb86b"} Oct 08 09:15:54 crc kubenswrapper[4744]: I1008 09:15:54.858012 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8s9r" event={"ID":"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4","Type":"ContainerStarted","Data":"db09d66d8bd75a67abdddc962a3dcf52ba2189b1c8168f70187c8e4e66e82e0c"} Oct 08 09:15:54 crc kubenswrapper[4744]: I1008 09:15:54.865249 4744 generic.go:334] "Generic (PLEG): container finished" podID="003a8292-d26b-47f8-89ce-60165fcad99e" containerID="bce56fcdbeee7e4654663d3c7b8de27c8fc8d39e95132b746121e80e3e8ee70c" exitCode=0 Oct 08 09:15:54 crc kubenswrapper[4744]: I1008 09:15:54.865347 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gqrp" event={"ID":"003a8292-d26b-47f8-89ce-60165fcad99e","Type":"ContainerDied","Data":"bce56fcdbeee7e4654663d3c7b8de27c8fc8d39e95132b746121e80e3e8ee70c"} Oct 08 09:15:54 crc kubenswrapper[4744]: I1008 09:15:54.865404 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gqrp" event={"ID":"003a8292-d26b-47f8-89ce-60165fcad99e","Type":"ContainerStarted","Data":"49fda692724e8db94088ec998930053c1646ec68f6d5c89f7caf6473ba304993"} Oct 08 09:15:55 crc kubenswrapper[4744]: I1008 09:15:55.880119 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q9562"] Oct 08 09:15:55 crc kubenswrapper[4744]: I1008 09:15:55.881310 4744 generic.go:334] "Generic (PLEG): container finished" podID="003a8292-d26b-47f8-89ce-60165fcad99e" containerID="fcf79a7e923c31ba4a6d55d63959577d6fa05c0e8efddb8586c5e37ad12d292b" exitCode=0 Oct 08 09:15:55 crc kubenswrapper[4744]: I1008 09:15:55.881532 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gqrp" event={"ID":"003a8292-d26b-47f8-89ce-60165fcad99e","Type":"ContainerDied","Data":"fcf79a7e923c31ba4a6d55d63959577d6fa05c0e8efddb8586c5e37ad12d292b"} Oct 08 09:15:55 crc kubenswrapper[4744]: I1008 09:15:55.881638 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:15:55 crc kubenswrapper[4744]: I1008 09:15:55.892783 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q9562"] Oct 08 09:15:55 crc kubenswrapper[4744]: I1008 09:15:55.892887 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.049676 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41990ef2-32ac-4477-a702-19e6fd228546-utilities\") pod \"redhat-operators-q9562\" (UID: \"41990ef2-32ac-4477-a702-19e6fd228546\") " pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.049728 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z67hj\" (UniqueName: \"kubernetes.io/projected/41990ef2-32ac-4477-a702-19e6fd228546-kube-api-access-z67hj\") pod \"redhat-operators-q9562\" (UID: \"41990ef2-32ac-4477-a702-19e6fd228546\") " pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.049817 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41990ef2-32ac-4477-a702-19e6fd228546-catalog-content\") pod \"redhat-operators-q9562\" (UID: \"41990ef2-32ac-4477-a702-19e6fd228546\") " pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.075354 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ffptv"] Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.093047 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ffptv"] Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.093162 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.095246 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.151008 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c528d4-e939-4c33-a145-cf8e58bb8c89-catalog-content\") pod \"community-operators-ffptv\" (UID: \"12c528d4-e939-4c33-a145-cf8e58bb8c89\") " pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.151076 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41990ef2-32ac-4477-a702-19e6fd228546-catalog-content\") pod \"redhat-operators-q9562\" (UID: \"41990ef2-32ac-4477-a702-19e6fd228546\") " pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.151103 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41990ef2-32ac-4477-a702-19e6fd228546-utilities\") pod \"redhat-operators-q9562\" (UID: \"41990ef2-32ac-4477-a702-19e6fd228546\") " pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.151123 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z67hj\" (UniqueName: \"kubernetes.io/projected/41990ef2-32ac-4477-a702-19e6fd228546-kube-api-access-z67hj\") pod \"redhat-operators-q9562\" (UID: \"41990ef2-32ac-4477-a702-19e6fd228546\") " pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.151141 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z45zh\" (UniqueName: \"kubernetes.io/projected/12c528d4-e939-4c33-a145-cf8e58bb8c89-kube-api-access-z45zh\") pod \"community-operators-ffptv\" (UID: \"12c528d4-e939-4c33-a145-cf8e58bb8c89\") " pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.151159 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c528d4-e939-4c33-a145-cf8e58bb8c89-utilities\") pod \"community-operators-ffptv\" (UID: \"12c528d4-e939-4c33-a145-cf8e58bb8c89\") " pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.151527 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41990ef2-32ac-4477-a702-19e6fd228546-catalog-content\") pod \"redhat-operators-q9562\" (UID: \"41990ef2-32ac-4477-a702-19e6fd228546\") " pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.151728 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41990ef2-32ac-4477-a702-19e6fd228546-utilities\") pod \"redhat-operators-q9562\" (UID: \"41990ef2-32ac-4477-a702-19e6fd228546\") " pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.185464 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z67hj\" (UniqueName: \"kubernetes.io/projected/41990ef2-32ac-4477-a702-19e6fd228546-kube-api-access-z67hj\") pod \"redhat-operators-q9562\" (UID: \"41990ef2-32ac-4477-a702-19e6fd228546\") " pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.213196 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.251917 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z45zh\" (UniqueName: \"kubernetes.io/projected/12c528d4-e939-4c33-a145-cf8e58bb8c89-kube-api-access-z45zh\") pod \"community-operators-ffptv\" (UID: \"12c528d4-e939-4c33-a145-cf8e58bb8c89\") " pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.252254 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c528d4-e939-4c33-a145-cf8e58bb8c89-utilities\") pod \"community-operators-ffptv\" (UID: \"12c528d4-e939-4c33-a145-cf8e58bb8c89\") " pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.252290 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c528d4-e939-4c33-a145-cf8e58bb8c89-catalog-content\") pod \"community-operators-ffptv\" (UID: \"12c528d4-e939-4c33-a145-cf8e58bb8c89\") " pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.252809 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c528d4-e939-4c33-a145-cf8e58bb8c89-catalog-content\") pod \"community-operators-ffptv\" (UID: \"12c528d4-e939-4c33-a145-cf8e58bb8c89\") " pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.253673 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c528d4-e939-4c33-a145-cf8e58bb8c89-utilities\") pod \"community-operators-ffptv\" (UID: \"12c528d4-e939-4c33-a145-cf8e58bb8c89\") " pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.300818 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z45zh\" (UniqueName: \"kubernetes.io/projected/12c528d4-e939-4c33-a145-cf8e58bb8c89-kube-api-access-z45zh\") pod \"community-operators-ffptv\" (UID: \"12c528d4-e939-4c33-a145-cf8e58bb8c89\") " pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.448928 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.601938 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q9562"] Oct 08 09:15:56 crc kubenswrapper[4744]: W1008 09:15:56.616520 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41990ef2_32ac_4477_a702_19e6fd228546.slice/crio-e53b2aafc717f5c8b234fb39ad504fb0f29eee18db7ad8fb2008aced43282a24 WatchSource:0}: Error finding container e53b2aafc717f5c8b234fb39ad504fb0f29eee18db7ad8fb2008aced43282a24: Status 404 returned error can't find the container with id e53b2aafc717f5c8b234fb39ad504fb0f29eee18db7ad8fb2008aced43282a24 Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.651860 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ffptv"] Oct 08 09:15:56 crc kubenswrapper[4744]: W1008 09:15:56.668767 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12c528d4_e939_4c33_a145_cf8e58bb8c89.slice/crio-51cef5ad9f435b683a102be9e91e5b3093f7c42eea80d8ecb99b49341ab26912 WatchSource:0}: Error finding container 51cef5ad9f435b683a102be9e91e5b3093f7c42eea80d8ecb99b49341ab26912: Status 404 returned error can't find the container with id 51cef5ad9f435b683a102be9e91e5b3093f7c42eea80d8ecb99b49341ab26912 Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.888965 4744 generic.go:334] "Generic (PLEG): container finished" podID="12c528d4-e939-4c33-a145-cf8e58bb8c89" containerID="43f7d2700ff61f54e19bd641d0076a3c7b63ba039e44e9f75a134f3e78ce9acc" exitCode=0 Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.889069 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ffptv" event={"ID":"12c528d4-e939-4c33-a145-cf8e58bb8c89","Type":"ContainerDied","Data":"43f7d2700ff61f54e19bd641d0076a3c7b63ba039e44e9f75a134f3e78ce9acc"} Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.889118 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ffptv" event={"ID":"12c528d4-e939-4c33-a145-cf8e58bb8c89","Type":"ContainerStarted","Data":"51cef5ad9f435b683a102be9e91e5b3093f7c42eea80d8ecb99b49341ab26912"} Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.893103 4744 generic.go:334] "Generic (PLEG): container finished" podID="8c8a122a-bcbf-474b-8c68-ed54aec6a6f4" containerID="cd5810b9c75330e0b33cd411dcb732562a6354ada74ad45c699f82b1d9f289e5" exitCode=0 Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.893199 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8s9r" event={"ID":"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4","Type":"ContainerDied","Data":"cd5810b9c75330e0b33cd411dcb732562a6354ada74ad45c699f82b1d9f289e5"} Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.894476 4744 generic.go:334] "Generic (PLEG): container finished" podID="41990ef2-32ac-4477-a702-19e6fd228546" containerID="8ef951c9670a310788341cfe15ca95de817d5c17eebea99063af514599a5dff9" exitCode=0 Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.894551 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9562" event={"ID":"41990ef2-32ac-4477-a702-19e6fd228546","Type":"ContainerDied","Data":"8ef951c9670a310788341cfe15ca95de817d5c17eebea99063af514599a5dff9"} Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.894570 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9562" event={"ID":"41990ef2-32ac-4477-a702-19e6fd228546","Type":"ContainerStarted","Data":"e53b2aafc717f5c8b234fb39ad504fb0f29eee18db7ad8fb2008aced43282a24"} Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.898741 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5gqrp" event={"ID":"003a8292-d26b-47f8-89ce-60165fcad99e","Type":"ContainerStarted","Data":"19dd75eec97c9bb990e1474456090fc3e47d17a34badf56f48f02daf00e7b95a"} Oct 08 09:15:56 crc kubenswrapper[4744]: I1008 09:15:56.963122 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5gqrp" podStartSLOduration=2.450543167 podStartE2EDuration="3.963102542s" podCreationTimestamp="2025-10-08 09:15:53 +0000 UTC" firstStartedPulling="2025-10-08 09:15:54.866685972 +0000 UTC m=+250.114331211" lastFinishedPulling="2025-10-08 09:15:56.379245347 +0000 UTC m=+251.626890586" observedRunningTime="2025-10-08 09:15:56.958486847 +0000 UTC m=+252.206132116" watchObservedRunningTime="2025-10-08 09:15:56.963102542 +0000 UTC m=+252.210747781" Oct 08 09:15:57 crc kubenswrapper[4744]: I1008 09:15:57.910907 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ffptv" event={"ID":"12c528d4-e939-4c33-a145-cf8e58bb8c89","Type":"ContainerStarted","Data":"c9b456199f4ec0e897493a53a14d4e4b8889349693cff0ea8cd3e38d514ac78b"} Oct 08 09:15:57 crc kubenswrapper[4744]: I1008 09:15:57.913832 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8s9r" event={"ID":"8c8a122a-bcbf-474b-8c68-ed54aec6a6f4","Type":"ContainerStarted","Data":"57cc992b1fe6e2330fb54af6e15205027c254119778998e428e021d2549a9e46"} Oct 08 09:15:57 crc kubenswrapper[4744]: I1008 09:15:57.952103 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g8s9r" podStartSLOduration=2.466764551 podStartE2EDuration="4.952083373s" podCreationTimestamp="2025-10-08 09:15:53 +0000 UTC" firstStartedPulling="2025-10-08 09:15:54.861527907 +0000 UTC m=+250.109173146" lastFinishedPulling="2025-10-08 09:15:57.346846719 +0000 UTC m=+252.594491968" observedRunningTime="2025-10-08 09:15:57.950708855 +0000 UTC m=+253.198354104" watchObservedRunningTime="2025-10-08 09:15:57.952083373 +0000 UTC m=+253.199728612" Oct 08 09:15:58 crc kubenswrapper[4744]: I1008 09:15:58.921679 4744 generic.go:334] "Generic (PLEG): container finished" podID="12c528d4-e939-4c33-a145-cf8e58bb8c89" containerID="c9b456199f4ec0e897493a53a14d4e4b8889349693cff0ea8cd3e38d514ac78b" exitCode=0 Oct 08 09:15:58 crc kubenswrapper[4744]: I1008 09:15:58.921829 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ffptv" event={"ID":"12c528d4-e939-4c33-a145-cf8e58bb8c89","Type":"ContainerDied","Data":"c9b456199f4ec0e897493a53a14d4e4b8889349693cff0ea8cd3e38d514ac78b"} Oct 08 09:15:58 crc kubenswrapper[4744]: I1008 09:15:58.922071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ffptv" event={"ID":"12c528d4-e939-4c33-a145-cf8e58bb8c89","Type":"ContainerStarted","Data":"740ecbe40a04d79d760e3a5b1f1c4ace223b08094638400c8049b4d3743200a4"} Oct 08 09:15:58 crc kubenswrapper[4744]: I1008 09:15:58.924262 4744 generic.go:334] "Generic (PLEG): container finished" podID="41990ef2-32ac-4477-a702-19e6fd228546" containerID="1e2cd455bfe6719acf70b3c9c98241c690a7ee952579eaa3f11691411a544854" exitCode=0 Oct 08 09:15:58 crc kubenswrapper[4744]: I1008 09:15:58.924317 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9562" event={"ID":"41990ef2-32ac-4477-a702-19e6fd228546","Type":"ContainerDied","Data":"1e2cd455bfe6719acf70b3c9c98241c690a7ee952579eaa3f11691411a544854"} Oct 08 09:15:58 crc kubenswrapper[4744]: I1008 09:15:58.941654 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ffptv" podStartSLOduration=1.534361378 podStartE2EDuration="2.941638786s" podCreationTimestamp="2025-10-08 09:15:56 +0000 UTC" firstStartedPulling="2025-10-08 09:15:56.895608894 +0000 UTC m=+252.143254133" lastFinishedPulling="2025-10-08 09:15:58.302886302 +0000 UTC m=+253.550531541" observedRunningTime="2025-10-08 09:15:58.94086434 +0000 UTC m=+254.188509589" watchObservedRunningTime="2025-10-08 09:15:58.941638786 +0000 UTC m=+254.189284015" Oct 08 09:16:00 crc kubenswrapper[4744]: I1008 09:16:00.943534 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9562" event={"ID":"41990ef2-32ac-4477-a702-19e6fd228546","Type":"ContainerStarted","Data":"612e0dfd3a5dc27d666dc6ec8ce67ccf959bde0fbefb5dd36ca526dc45d3285c"} Oct 08 09:16:00 crc kubenswrapper[4744]: I1008 09:16:00.963674 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q9562" podStartSLOduration=3.183153959 podStartE2EDuration="5.963653225s" podCreationTimestamp="2025-10-08 09:15:55 +0000 UTC" firstStartedPulling="2025-10-08 09:15:56.895620784 +0000 UTC m=+252.143266023" lastFinishedPulling="2025-10-08 09:15:59.67612005 +0000 UTC m=+254.923765289" observedRunningTime="2025-10-08 09:16:00.961703474 +0000 UTC m=+256.209348753" watchObservedRunningTime="2025-10-08 09:16:00.963653225 +0000 UTC m=+256.211298474" Oct 08 09:16:03 crc kubenswrapper[4744]: I1008 09:16:03.809138 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:16:03 crc kubenswrapper[4744]: I1008 09:16:03.809621 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:16:03 crc kubenswrapper[4744]: I1008 09:16:03.857428 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:16:03 crc kubenswrapper[4744]: I1008 09:16:03.992527 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g8s9r" Oct 08 09:16:04 crc kubenswrapper[4744]: I1008 09:16:04.001952 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:16:04 crc kubenswrapper[4744]: I1008 09:16:04.002065 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:16:04 crc kubenswrapper[4744]: I1008 09:16:04.089994 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:16:05 crc kubenswrapper[4744]: I1008 09:16:05.000456 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5gqrp" Oct 08 09:16:06 crc kubenswrapper[4744]: I1008 09:16:06.214031 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:16:06 crc kubenswrapper[4744]: I1008 09:16:06.214468 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:16:06 crc kubenswrapper[4744]: I1008 09:16:06.253963 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:16:06 crc kubenswrapper[4744]: I1008 09:16:06.449468 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:16:06 crc kubenswrapper[4744]: I1008 09:16:06.449525 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:16:06 crc kubenswrapper[4744]: I1008 09:16:06.507957 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:16:07 crc kubenswrapper[4744]: I1008 09:16:07.008525 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q9562" Oct 08 09:16:07 crc kubenswrapper[4744]: I1008 09:16:07.009882 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ffptv" Oct 08 09:17:19 crc kubenswrapper[4744]: I1008 09:17:19.689868 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:17:19 crc kubenswrapper[4744]: I1008 09:17:19.690453 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:17:49 crc kubenswrapper[4744]: I1008 09:17:49.691000 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:17:49 crc kubenswrapper[4744]: I1008 09:17:49.692231 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:18:19 crc kubenswrapper[4744]: I1008 09:18:19.690712 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:18:19 crc kubenswrapper[4744]: I1008 09:18:19.691234 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:18:19 crc kubenswrapper[4744]: I1008 09:18:19.691291 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:18:19 crc kubenswrapper[4744]: I1008 09:18:19.692000 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"63e1dfa02860e1e750f72656194fd55b1911a02f76a9cd7c271f9f20e1dffa0a"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 09:18:19 crc kubenswrapper[4744]: I1008 09:18:19.692075 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://63e1dfa02860e1e750f72656194fd55b1911a02f76a9cd7c271f9f20e1dffa0a" gracePeriod=600 Oct 08 09:18:19 crc kubenswrapper[4744]: I1008 09:18:19.894899 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="63e1dfa02860e1e750f72656194fd55b1911a02f76a9cd7c271f9f20e1dffa0a" exitCode=0 Oct 08 09:18:19 crc kubenswrapper[4744]: I1008 09:18:19.894999 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"63e1dfa02860e1e750f72656194fd55b1911a02f76a9cd7c271f9f20e1dffa0a"} Oct 08 09:18:19 crc kubenswrapper[4744]: I1008 09:18:19.895405 4744 scope.go:117] "RemoveContainer" containerID="3220c989105b826ade7af9e7b313b256db0f3069393b20ddf588de89c5163cf0" Oct 08 09:18:20 crc kubenswrapper[4744]: I1008 09:18:20.908526 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"8c068a91bb7d0e26249b43fab5fe60ecb8e806b9d40583c0517d3afb38ddf715"} Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.008124 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2h9gn"] Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.009751 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.077685 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e655ef7f-71cd-43ef-8900-e7b0e08c4806-registry-certificates\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.077753 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e655ef7f-71cd-43ef-8900-e7b0e08c4806-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.077780 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e655ef7f-71cd-43ef-8900-e7b0e08c4806-registry-tls\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.077810 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.077835 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnknp\" (UniqueName: \"kubernetes.io/projected/e655ef7f-71cd-43ef-8900-e7b0e08c4806-kube-api-access-rnknp\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.077858 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e655ef7f-71cd-43ef-8900-e7b0e08c4806-bound-sa-token\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.077874 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e655ef7f-71cd-43ef-8900-e7b0e08c4806-trusted-ca\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.077891 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e655ef7f-71cd-43ef-8900-e7b0e08c4806-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.092727 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2h9gn"] Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.105938 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.178695 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e655ef7f-71cd-43ef-8900-e7b0e08c4806-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.178745 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e655ef7f-71cd-43ef-8900-e7b0e08c4806-registry-tls\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.178785 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnknp\" (UniqueName: \"kubernetes.io/projected/e655ef7f-71cd-43ef-8900-e7b0e08c4806-kube-api-access-rnknp\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.178809 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e655ef7f-71cd-43ef-8900-e7b0e08c4806-bound-sa-token\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.178830 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e655ef7f-71cd-43ef-8900-e7b0e08c4806-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.178847 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e655ef7f-71cd-43ef-8900-e7b0e08c4806-trusted-ca\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.178869 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e655ef7f-71cd-43ef-8900-e7b0e08c4806-registry-certificates\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.180096 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e655ef7f-71cd-43ef-8900-e7b0e08c4806-ca-trust-extracted\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.180658 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e655ef7f-71cd-43ef-8900-e7b0e08c4806-registry-certificates\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.181034 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e655ef7f-71cd-43ef-8900-e7b0e08c4806-trusted-ca\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.184993 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e655ef7f-71cd-43ef-8900-e7b0e08c4806-registry-tls\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.184998 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e655ef7f-71cd-43ef-8900-e7b0e08c4806-installation-pull-secrets\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.196468 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e655ef7f-71cd-43ef-8900-e7b0e08c4806-bound-sa-token\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.196724 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnknp\" (UniqueName: \"kubernetes.io/projected/e655ef7f-71cd-43ef-8900-e7b0e08c4806-kube-api-access-rnknp\") pod \"image-registry-66df7c8f76-2h9gn\" (UID: \"e655ef7f-71cd-43ef-8900-e7b0e08c4806\") " pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.328682 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.532954 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-2h9gn"] Oct 08 09:20:10 crc kubenswrapper[4744]: W1008 09:20:10.538737 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode655ef7f_71cd_43ef_8900_e7b0e08c4806.slice/crio-734d321897465c6728fbd51a8e51e93781881c553d41e01cc9873cc011e3dd11 WatchSource:0}: Error finding container 734d321897465c6728fbd51a8e51e93781881c553d41e01cc9873cc011e3dd11: Status 404 returned error can't find the container with id 734d321897465c6728fbd51a8e51e93781881c553d41e01cc9873cc011e3dd11 Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.634740 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" event={"ID":"e655ef7f-71cd-43ef-8900-e7b0e08c4806","Type":"ContainerStarted","Data":"31dd802b7e245af0c3d23f1de182b7204778398250495135fa124da4c2740655"} Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.635090 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" event={"ID":"e655ef7f-71cd-43ef-8900-e7b0e08c4806","Type":"ContainerStarted","Data":"734d321897465c6728fbd51a8e51e93781881c553d41e01cc9873cc011e3dd11"} Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.635130 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:10 crc kubenswrapper[4744]: I1008 09:20:10.659731 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" podStartSLOduration=1.659709115 podStartE2EDuration="1.659709115s" podCreationTimestamp="2025-10-08 09:20:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:20:10.652354042 +0000 UTC m=+505.899999311" watchObservedRunningTime="2025-10-08 09:20:10.659709115 +0000 UTC m=+505.907354354" Oct 08 09:20:19 crc kubenswrapper[4744]: I1008 09:20:19.690731 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:20:19 crc kubenswrapper[4744]: I1008 09:20:19.691566 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:20:30 crc kubenswrapper[4744]: I1008 09:20:30.344255 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-2h9gn" Oct 08 09:20:30 crc kubenswrapper[4744]: I1008 09:20:30.435311 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5lxjl"] Oct 08 09:20:49 crc kubenswrapper[4744]: I1008 09:20:49.689783 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:20:49 crc kubenswrapper[4744]: I1008 09:20:49.690303 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.485336 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" podUID="1b2217f7-f478-4833-b301-097e7c1c07c9" containerName="registry" containerID="cri-o://565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb" gracePeriod=30 Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.806266 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.896788 4744 generic.go:334] "Generic (PLEG): container finished" podID="1b2217f7-f478-4833-b301-097e7c1c07c9" containerID="565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb" exitCode=0 Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.896833 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.896847 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" event={"ID":"1b2217f7-f478-4833-b301-097e7c1c07c9","Type":"ContainerDied","Data":"565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb"} Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.896900 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5lxjl" event={"ID":"1b2217f7-f478-4833-b301-097e7c1c07c9","Type":"ContainerDied","Data":"931e961dbdc35ba5c0be812dc9fd984ea1eb77eaae512274d70b032eb2dea1f0"} Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.896921 4744 scope.go:117] "RemoveContainer" containerID="565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.911177 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2hpx\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-kube-api-access-s2hpx\") pod \"1b2217f7-f478-4833-b301-097e7c1c07c9\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.911222 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b2217f7-f478-4833-b301-097e7c1c07c9-trusted-ca\") pod \"1b2217f7-f478-4833-b301-097e7c1c07c9\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.911250 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-bound-sa-token\") pod \"1b2217f7-f478-4833-b301-097e7c1c07c9\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.911302 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1b2217f7-f478-4833-b301-097e7c1c07c9-ca-trust-extracted\") pod \"1b2217f7-f478-4833-b301-097e7c1c07c9\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.911478 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"1b2217f7-f478-4833-b301-097e7c1c07c9\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.911516 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-registry-tls\") pod \"1b2217f7-f478-4833-b301-097e7c1c07c9\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.911543 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1b2217f7-f478-4833-b301-097e7c1c07c9-registry-certificates\") pod \"1b2217f7-f478-4833-b301-097e7c1c07c9\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.911569 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1b2217f7-f478-4833-b301-097e7c1c07c9-installation-pull-secrets\") pod \"1b2217f7-f478-4833-b301-097e7c1c07c9\" (UID: \"1b2217f7-f478-4833-b301-097e7c1c07c9\") " Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.912063 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b2217f7-f478-4833-b301-097e7c1c07c9-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "1b2217f7-f478-4833-b301-097e7c1c07c9" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.913657 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b2217f7-f478-4833-b301-097e7c1c07c9-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "1b2217f7-f478-4833-b301-097e7c1c07c9" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.919405 4744 scope.go:117] "RemoveContainer" containerID="565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.920691 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-kube-api-access-s2hpx" (OuterVolumeSpecName: "kube-api-access-s2hpx") pod "1b2217f7-f478-4833-b301-097e7c1c07c9" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9"). InnerVolumeSpecName "kube-api-access-s2hpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:20:55 crc kubenswrapper[4744]: E1008 09:20:55.921109 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb\": container with ID starting with 565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb not found: ID does not exist" containerID="565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.921141 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb"} err="failed to get container status \"565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb\": rpc error: code = NotFound desc = could not find container \"565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb\": container with ID starting with 565fba62099a9cb6338deae51d0db9730f02490c0ae9b5abc88fbb9e478a4cfb not found: ID does not exist" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.925964 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "1b2217f7-f478-4833-b301-097e7c1c07c9" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.926352 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2217f7-f478-4833-b301-097e7c1c07c9-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "1b2217f7-f478-4833-b301-097e7c1c07c9" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.927865 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "1b2217f7-f478-4833-b301-097e7c1c07c9" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.930355 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "1b2217f7-f478-4833-b301-097e7c1c07c9" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:20:55 crc kubenswrapper[4744]: I1008 09:20:55.930489 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b2217f7-f478-4833-b301-097e7c1c07c9-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "1b2217f7-f478-4833-b301-097e7c1c07c9" (UID: "1b2217f7-f478-4833-b301-097e7c1c07c9"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:20:56 crc kubenswrapper[4744]: I1008 09:20:56.012705 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2hpx\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-kube-api-access-s2hpx\") on node \"crc\" DevicePath \"\"" Oct 08 09:20:56 crc kubenswrapper[4744]: I1008 09:20:56.012736 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1b2217f7-f478-4833-b301-097e7c1c07c9-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:20:56 crc kubenswrapper[4744]: I1008 09:20:56.012749 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 09:20:56 crc kubenswrapper[4744]: I1008 09:20:56.012757 4744 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1b2217f7-f478-4833-b301-097e7c1c07c9-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 09:20:56 crc kubenswrapper[4744]: I1008 09:20:56.012765 4744 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1b2217f7-f478-4833-b301-097e7c1c07c9-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:20:56 crc kubenswrapper[4744]: I1008 09:20:56.012773 4744 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1b2217f7-f478-4833-b301-097e7c1c07c9-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 09:20:56 crc kubenswrapper[4744]: I1008 09:20:56.012784 4744 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1b2217f7-f478-4833-b301-097e7c1c07c9-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 09:20:56 crc kubenswrapper[4744]: I1008 09:20:56.247401 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5lxjl"] Oct 08 09:20:56 crc kubenswrapper[4744]: I1008 09:20:56.251625 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5lxjl"] Oct 08 09:20:57 crc kubenswrapper[4744]: I1008 09:20:57.463432 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b2217f7-f478-4833-b301-097e7c1c07c9" path="/var/lib/kubelet/pods/1b2217f7-f478-4833-b301-097e7c1c07c9/volumes" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.358857 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-j5pzr"] Oct 08 09:20:58 crc kubenswrapper[4744]: E1008 09:20:58.359102 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2217f7-f478-4833-b301-097e7c1c07c9" containerName="registry" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.359124 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2217f7-f478-4833-b301-097e7c1c07c9" containerName="registry" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.359256 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2217f7-f478-4833-b301-097e7c1c07c9" containerName="registry" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.359741 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-j5pzr" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.362077 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.362159 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.362077 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-tsxjv" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.372018 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-87br5"] Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.372823 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-87br5" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.376317 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-qqdhs" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.389217 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4478x"] Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.390162 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-4478x" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.393108 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-87dmh" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.395138 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-87br5"] Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.411484 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4478x"] Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.419636 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-j5pzr"] Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.547208 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lcg9\" (UniqueName: \"kubernetes.io/projected/39d1521a-c337-4259-bc31-0bf6357ce325-kube-api-access-5lcg9\") pod \"cert-manager-webhook-5655c58dd6-4478x\" (UID: \"39d1521a-c337-4259-bc31-0bf6357ce325\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4478x" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.547275 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k84d\" (UniqueName: \"kubernetes.io/projected/b2fe7e00-2fdb-4186-81db-c211ef88b772-kube-api-access-9k84d\") pod \"cert-manager-cainjector-7f985d654d-j5pzr\" (UID: \"b2fe7e00-2fdb-4186-81db-c211ef88b772\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-j5pzr" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.547295 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwcrb\" (UniqueName: \"kubernetes.io/projected/5fde9304-c103-4e8a-96bc-d47a0dea1635-kube-api-access-fwcrb\") pod \"cert-manager-5b446d88c5-87br5\" (UID: \"5fde9304-c103-4e8a-96bc-d47a0dea1635\") " pod="cert-manager/cert-manager-5b446d88c5-87br5" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.647815 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k84d\" (UniqueName: \"kubernetes.io/projected/b2fe7e00-2fdb-4186-81db-c211ef88b772-kube-api-access-9k84d\") pod \"cert-manager-cainjector-7f985d654d-j5pzr\" (UID: \"b2fe7e00-2fdb-4186-81db-c211ef88b772\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-j5pzr" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.647873 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwcrb\" (UniqueName: \"kubernetes.io/projected/5fde9304-c103-4e8a-96bc-d47a0dea1635-kube-api-access-fwcrb\") pod \"cert-manager-5b446d88c5-87br5\" (UID: \"5fde9304-c103-4e8a-96bc-d47a0dea1635\") " pod="cert-manager/cert-manager-5b446d88c5-87br5" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.647936 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lcg9\" (UniqueName: \"kubernetes.io/projected/39d1521a-c337-4259-bc31-0bf6357ce325-kube-api-access-5lcg9\") pod \"cert-manager-webhook-5655c58dd6-4478x\" (UID: \"39d1521a-c337-4259-bc31-0bf6357ce325\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4478x" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.666832 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k84d\" (UniqueName: \"kubernetes.io/projected/b2fe7e00-2fdb-4186-81db-c211ef88b772-kube-api-access-9k84d\") pod \"cert-manager-cainjector-7f985d654d-j5pzr\" (UID: \"b2fe7e00-2fdb-4186-81db-c211ef88b772\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-j5pzr" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.667883 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwcrb\" (UniqueName: \"kubernetes.io/projected/5fde9304-c103-4e8a-96bc-d47a0dea1635-kube-api-access-fwcrb\") pod \"cert-manager-5b446d88c5-87br5\" (UID: \"5fde9304-c103-4e8a-96bc-d47a0dea1635\") " pod="cert-manager/cert-manager-5b446d88c5-87br5" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.668807 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lcg9\" (UniqueName: \"kubernetes.io/projected/39d1521a-c337-4259-bc31-0bf6357ce325-kube-api-access-5lcg9\") pod \"cert-manager-webhook-5655c58dd6-4478x\" (UID: \"39d1521a-c337-4259-bc31-0bf6357ce325\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-4478x" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.672545 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-j5pzr" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.686614 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-87br5" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.702599 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-4478x" Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.879514 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-j5pzr"] Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.894308 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 09:20:58 crc kubenswrapper[4744]: I1008 09:20:58.929959 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-j5pzr" event={"ID":"b2fe7e00-2fdb-4186-81db-c211ef88b772","Type":"ContainerStarted","Data":"e1d49f3e6314da6811e6f4bd4decd0a0b1d0875e79e772adb09e63b5281fee8d"} Oct 08 09:20:59 crc kubenswrapper[4744]: I1008 09:20:59.161324 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-4478x"] Oct 08 09:20:59 crc kubenswrapper[4744]: W1008 09:20:59.166265 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39d1521a_c337_4259_bc31_0bf6357ce325.slice/crio-99d18bb48bf36784b88bb9c86fef37b9d533496cd8eaff1bb55f1bce329031f4 WatchSource:0}: Error finding container 99d18bb48bf36784b88bb9c86fef37b9d533496cd8eaff1bb55f1bce329031f4: Status 404 returned error can't find the container with id 99d18bb48bf36784b88bb9c86fef37b9d533496cd8eaff1bb55f1bce329031f4 Oct 08 09:20:59 crc kubenswrapper[4744]: I1008 09:20:59.172616 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-87br5"] Oct 08 09:20:59 crc kubenswrapper[4744]: W1008 09:20:59.180169 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fde9304_c103_4e8a_96bc_d47a0dea1635.slice/crio-fd4e6bf9685008376b685855cec8fbfb5a8c3c707787a56635f238aa009989b2 WatchSource:0}: Error finding container fd4e6bf9685008376b685855cec8fbfb5a8c3c707787a56635f238aa009989b2: Status 404 returned error can't find the container with id fd4e6bf9685008376b685855cec8fbfb5a8c3c707787a56635f238aa009989b2 Oct 08 09:20:59 crc kubenswrapper[4744]: I1008 09:20:59.937635 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-4478x" event={"ID":"39d1521a-c337-4259-bc31-0bf6357ce325","Type":"ContainerStarted","Data":"99d18bb48bf36784b88bb9c86fef37b9d533496cd8eaff1bb55f1bce329031f4"} Oct 08 09:20:59 crc kubenswrapper[4744]: I1008 09:20:59.938708 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-87br5" event={"ID":"5fde9304-c103-4e8a-96bc-d47a0dea1635","Type":"ContainerStarted","Data":"fd4e6bf9685008376b685855cec8fbfb5a8c3c707787a56635f238aa009989b2"} Oct 08 09:21:00 crc kubenswrapper[4744]: I1008 09:21:00.945705 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-j5pzr" event={"ID":"b2fe7e00-2fdb-4186-81db-c211ef88b772","Type":"ContainerStarted","Data":"74928efa1bc0012e3d7d2789c46ecd8e115b495562cbb59b916c81e30678ed63"} Oct 08 09:21:00 crc kubenswrapper[4744]: I1008 09:21:00.962319 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-j5pzr" podStartSLOduration=1.166828032 podStartE2EDuration="2.962257685s" podCreationTimestamp="2025-10-08 09:20:58 +0000 UTC" firstStartedPulling="2025-10-08 09:20:58.894047493 +0000 UTC m=+554.141692732" lastFinishedPulling="2025-10-08 09:21:00.689477146 +0000 UTC m=+555.937122385" observedRunningTime="2025-10-08 09:21:00.961318829 +0000 UTC m=+556.208964068" watchObservedRunningTime="2025-10-08 09:21:00.962257685 +0000 UTC m=+556.209902924" Oct 08 09:21:02 crc kubenswrapper[4744]: I1008 09:21:02.962011 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-4478x" event={"ID":"39d1521a-c337-4259-bc31-0bf6357ce325","Type":"ContainerStarted","Data":"112c50e53990a327015feaf68fdd151a798cd11213ae7b157c68e9a134c563e1"} Oct 08 09:21:02 crc kubenswrapper[4744]: I1008 09:21:02.963465 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-4478x" Oct 08 09:21:02 crc kubenswrapper[4744]: I1008 09:21:02.965100 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-87br5" event={"ID":"5fde9304-c103-4e8a-96bc-d47a0dea1635","Type":"ContainerStarted","Data":"0399b46a4f5812f142084f53ffa4885d986fc025e7e11af3e66142dcb0e84d51"} Oct 08 09:21:03 crc kubenswrapper[4744]: I1008 09:21:03.011526 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-4478x" podStartSLOduration=2.019404333 podStartE2EDuration="5.01150546s" podCreationTimestamp="2025-10-08 09:20:58 +0000 UTC" firstStartedPulling="2025-10-08 09:20:59.168288515 +0000 UTC m=+554.415933754" lastFinishedPulling="2025-10-08 09:21:02.160389642 +0000 UTC m=+557.408034881" observedRunningTime="2025-10-08 09:21:02.990706877 +0000 UTC m=+558.238352136" watchObservedRunningTime="2025-10-08 09:21:03.01150546 +0000 UTC m=+558.259150699" Oct 08 09:21:03 crc kubenswrapper[4744]: I1008 09:21:03.012533 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-87br5" podStartSLOduration=2.097639508 podStartE2EDuration="5.012524619s" podCreationTimestamp="2025-10-08 09:20:58 +0000 UTC" firstStartedPulling="2025-10-08 09:20:59.182393673 +0000 UTC m=+554.430038912" lastFinishedPulling="2025-10-08 09:21:02.097278774 +0000 UTC m=+557.344924023" observedRunningTime="2025-10-08 09:21:03.006903876 +0000 UTC m=+558.254549115" watchObservedRunningTime="2025-10-08 09:21:03.012524619 +0000 UTC m=+558.260169858" Oct 08 09:21:08 crc kubenswrapper[4744]: I1008 09:21:08.707997 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-4478x" Oct 08 09:21:08 crc kubenswrapper[4744]: I1008 09:21:08.730985 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tmmnj"] Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.009117 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovn-controller" containerID="cri-o://bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe" gracePeriod=30 Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.009550 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="sbdb" containerID="cri-o://7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e" gracePeriod=30 Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.009601 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="nbdb" containerID="cri-o://31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e" gracePeriod=30 Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.009637 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="northd" containerID="cri-o://411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad" gracePeriod=30 Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.009677 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72" gracePeriod=30 Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.009713 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="kube-rbac-proxy-node" containerID="cri-o://3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb" gracePeriod=30 Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.009772 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovn-acl-logging" containerID="cri-o://0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1" gracePeriod=30 Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.065709 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" containerID="cri-o://db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a" gracePeriod=30 Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.355074 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/3.log" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.359072 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovn-acl-logging/0.log" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.359944 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovn-controller/0.log" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.361004 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411220 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovn-node-metrics-cert\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411312 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovnkube-script-lib\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411345 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-run-ovn-kubernetes\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411360 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-var-lib-openvswitch\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411394 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-systemd-units\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411431 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-slash\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411485 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-ovn\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411513 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-openvswitch\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411542 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-log-socket\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411576 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q28tm\" (UniqueName: \"kubernetes.io/projected/5a9bf215-f18b-4732-b12c-0355a3a0a15f-kube-api-access-q28tm\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411596 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-node-log\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411631 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovnkube-config\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411667 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-systemd\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411703 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-run-netns\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411719 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-kubelet\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411734 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411755 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-cni-netd\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411774 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-etc-openvswitch\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411794 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-env-overrides\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.411815 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-cni-bin\") pod \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\" (UID: \"5a9bf215-f18b-4732-b12c-0355a3a0a15f\") " Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412167 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412196 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cjbt2"] Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412445 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412466 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412477 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412485 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412496 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412502 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412512 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="sbdb" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412519 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="sbdb" Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412530 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412538 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412554 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="kube-rbac-proxy-node" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412561 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="kube-rbac-proxy-node" Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412571 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovn-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412578 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovn-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412585 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="northd" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412592 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="northd" Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412601 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="nbdb" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412608 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="nbdb" Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412617 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412623 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412634 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovn-acl-logging" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412641 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovn-acl-logging" Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412649 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="kubecfg-setup" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412657 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="kubecfg-setup" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412766 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="sbdb" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412779 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovn-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412788 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412796 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="northd" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412807 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412815 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="kube-rbac-proxy-node" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412824 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="nbdb" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412834 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovn-acl-logging" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412844 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412851 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412859 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 09:21:09 crc kubenswrapper[4744]: E1008 09:21:09.412961 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.412968 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.413098 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerName="ovnkube-controller" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.414424 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.414672 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-node-log" (OuterVolumeSpecName: "node-log") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.414962 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.414998 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.415064 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.415444 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.415950 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.415483 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.415511 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.415511 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.415535 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.415533 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.415552 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.415563 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-log-socket" (OuterVolumeSpecName: "log-socket") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.415595 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-slash" (OuterVolumeSpecName: "host-slash") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.415961 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.414993 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.425299 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.428887 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a9bf215-f18b-4732-b12c-0355a3a0a15f-kube-api-access-q28tm" (OuterVolumeSpecName: "kube-api-access-q28tm") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "kube-api-access-q28tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.438505 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "5a9bf215-f18b-4732-b12c-0355a3a0a15f" (UID: "5a9bf215-f18b-4732-b12c-0355a3a0a15f"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.513909 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1eb3be1b-60dd-44d8-b727-1cab861c60e5-ovnkube-script-lib\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.513988 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-run-openvswitch\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.514014 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-kubelet\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.514196 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-log-socket\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.514262 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-run-ovn\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.514287 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-run-systemd\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.514317 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-var-lib-openvswitch\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.514598 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-run-ovn-kubernetes\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.514813 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-node-log\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.514912 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1eb3be1b-60dd-44d8-b727-1cab861c60e5-env-overrides\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.514977 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-run-netns\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.515060 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1eb3be1b-60dd-44d8-b727-1cab861c60e5-ovnkube-config\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.515132 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqlkc\" (UniqueName: \"kubernetes.io/projected/1eb3be1b-60dd-44d8-b727-1cab861c60e5-kube-api-access-lqlkc\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.515268 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1eb3be1b-60dd-44d8-b727-1cab861c60e5-ovn-node-metrics-cert\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.515344 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-systemd-units\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.515489 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-cni-bin\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.515556 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-cni-netd\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.515632 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-slash\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.515679 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.515754 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-etc-openvswitch\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.515976 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516002 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516020 4744 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516032 4744 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516122 4744 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516136 4744 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-slash\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516150 4744 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516161 4744 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516180 4744 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-log-socket\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516193 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q28tm\" (UniqueName: \"kubernetes.io/projected/5a9bf215-f18b-4732-b12c-0355a3a0a15f-kube-api-access-q28tm\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516205 4744 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-node-log\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516217 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516228 4744 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516238 4744 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516249 4744 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516263 4744 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516276 4744 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516288 4744 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516300 4744 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a9bf215-f18b-4732-b12c-0355a3a0a15f-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.516313 4744 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5a9bf215-f18b-4732-b12c-0355a3a0a15f-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.617609 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1eb3be1b-60dd-44d8-b727-1cab861c60e5-ovn-node-metrics-cert\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.617703 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-systemd-units\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.617768 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-cni-bin\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.617798 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-cni-netd\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.617835 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-slash\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.617883 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.617887 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-cni-bin\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.617937 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-slash\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.617968 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.617949 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-etc-openvswitch\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.617909 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-cni-netd\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618007 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-etc-openvswitch\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618052 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-systemd-units\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618170 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1eb3be1b-60dd-44d8-b727-1cab861c60e5-ovnkube-script-lib\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618223 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-kubelet\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-run-openvswitch\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618291 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-log-socket\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618328 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-run-ovn\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618353 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-run-systemd\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618415 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-var-lib-openvswitch\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618448 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-run-ovn-kubernetes\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618518 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-node-log\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618548 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-run-openvswitch\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618568 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-kubelet\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618596 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-run-ovn-kubernetes\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618492 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-log-socket\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618554 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1eb3be1b-60dd-44d8-b727-1cab861c60e5-env-overrides\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618639 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-run-ovn\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618637 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-node-log\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618599 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-var-lib-openvswitch\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618677 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-run-netns\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618700 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-host-run-netns\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618704 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1eb3be1b-60dd-44d8-b727-1cab861c60e5-ovnkube-config\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618733 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqlkc\" (UniqueName: \"kubernetes.io/projected/1eb3be1b-60dd-44d8-b727-1cab861c60e5-kube-api-access-lqlkc\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.618544 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1eb3be1b-60dd-44d8-b727-1cab861c60e5-run-systemd\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.619323 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1eb3be1b-60dd-44d8-b727-1cab861c60e5-env-overrides\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.619350 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1eb3be1b-60dd-44d8-b727-1cab861c60e5-ovnkube-script-lib\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.619805 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1eb3be1b-60dd-44d8-b727-1cab861c60e5-ovnkube-config\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.623956 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1eb3be1b-60dd-44d8-b727-1cab861c60e5-ovn-node-metrics-cert\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.637280 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqlkc\" (UniqueName: \"kubernetes.io/projected/1eb3be1b-60dd-44d8-b727-1cab861c60e5-kube-api-access-lqlkc\") pod \"ovnkube-node-cjbt2\" (UID: \"1eb3be1b-60dd-44d8-b727-1cab861c60e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: I1008 09:21:09.740855 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:09 crc kubenswrapper[4744]: W1008 09:21:09.787488 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1eb3be1b_60dd_44d8_b727_1cab861c60e5.slice/crio-c5e1b7efb869b0fc807837c8578b9ea9e75f5f16da9a0a0707bc4d8249fc3719 WatchSource:0}: Error finding container c5e1b7efb869b0fc807837c8578b9ea9e75f5f16da9a0a0707bc4d8249fc3719: Status 404 returned error can't find the container with id c5e1b7efb869b0fc807837c8578b9ea9e75f5f16da9a0a0707bc4d8249fc3719 Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.017323 4744 generic.go:334] "Generic (PLEG): container finished" podID="1eb3be1b-60dd-44d8-b727-1cab861c60e5" containerID="56d477bdee2cfe8ccbd20bf4e3fb28e8c5988a9e3011b24c525a7d0800ba762b" exitCode=0 Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.017444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" event={"ID":"1eb3be1b-60dd-44d8-b727-1cab861c60e5","Type":"ContainerDied","Data":"56d477bdee2cfe8ccbd20bf4e3fb28e8c5988a9e3011b24c525a7d0800ba762b"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.017489 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" event={"ID":"1eb3be1b-60dd-44d8-b727-1cab861c60e5","Type":"ContainerStarted","Data":"c5e1b7efb869b0fc807837c8578b9ea9e75f5f16da9a0a0707bc4d8249fc3719"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.023007 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovnkube-controller/3.log" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.026721 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovn-acl-logging/0.log" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.027334 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tmmnj_5a9bf215-f18b-4732-b12c-0355a3a0a15f/ovn-controller/0.log" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.027974 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a" exitCode=0 Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028022 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e" exitCode=0 Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028041 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e" exitCode=0 Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028055 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad" exitCode=0 Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028069 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72" exitCode=0 Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028083 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb" exitCode=0 Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028096 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1" exitCode=143 Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028110 4744 generic.go:334] "Generic (PLEG): container finished" podID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" containerID="bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe" exitCode=143 Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028209 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028317 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028352 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028408 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028434 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028460 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028491 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028516 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028530 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028545 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028557 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028567 4744 scope.go:117] "RemoveContainer" containerID="db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028559 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028670 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028686 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028702 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028716 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028740 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028765 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028781 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028795 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028809 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028825 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028838 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028850 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028864 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028877 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028891 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028910 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028931 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028947 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028962 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028976 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.028990 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029004 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029017 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029030 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029042 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029055 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029075 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tmmnj" event={"ID":"5a9bf215-f18b-4732-b12c-0355a3a0a15f","Type":"ContainerDied","Data":"bda60c6f4059e554ef5635a19438017eaf5b62aa6fadef4522cdcf59850de381"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029096 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029111 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029124 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029137 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029151 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029165 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029180 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029214 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029230 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.029243 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.033836 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dldlq_7ea48507-1f33-46fc-ab34-c6e083dd86a9/kube-multus/2.log" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.034490 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dldlq_7ea48507-1f33-46fc-ab34-c6e083dd86a9/kube-multus/1.log" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.034541 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ea48507-1f33-46fc-ab34-c6e083dd86a9" containerID="94b68c9a4976717260751eead14fea87839d1758f32a92629bc6776f522afb3e" exitCode=2 Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.034579 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dldlq" event={"ID":"7ea48507-1f33-46fc-ab34-c6e083dd86a9","Type":"ContainerDied","Data":"94b68c9a4976717260751eead14fea87839d1758f32a92629bc6776f522afb3e"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.034612 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75"} Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.035042 4744 scope.go:117] "RemoveContainer" containerID="94b68c9a4976717260751eead14fea87839d1758f32a92629bc6776f522afb3e" Oct 08 09:21:10 crc kubenswrapper[4744]: E1008 09:21:10.035235 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-dldlq_openshift-multus(7ea48507-1f33-46fc-ab34-c6e083dd86a9)\"" pod="openshift-multus/multus-dldlq" podUID="7ea48507-1f33-46fc-ab34-c6e083dd86a9" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.062493 4744 scope.go:117] "RemoveContainer" containerID="00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.103894 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tmmnj"] Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.116434 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tmmnj"] Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.120659 4744 scope.go:117] "RemoveContainer" containerID="7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.174282 4744 scope.go:117] "RemoveContainer" containerID="31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.193283 4744 scope.go:117] "RemoveContainer" containerID="411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.240595 4744 scope.go:117] "RemoveContainer" containerID="5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.259224 4744 scope.go:117] "RemoveContainer" containerID="3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.276293 4744 scope.go:117] "RemoveContainer" containerID="0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.296514 4744 scope.go:117] "RemoveContainer" containerID="bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.315322 4744 scope.go:117] "RemoveContainer" containerID="25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.346054 4744 scope.go:117] "RemoveContainer" containerID="db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a" Oct 08 09:21:10 crc kubenswrapper[4744]: E1008 09:21:10.346767 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a\": container with ID starting with db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a not found: ID does not exist" containerID="db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.346818 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a"} err="failed to get container status \"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a\": rpc error: code = NotFound desc = could not find container \"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a\": container with ID starting with db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.346852 4744 scope.go:117] "RemoveContainer" containerID="00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357" Oct 08 09:21:10 crc kubenswrapper[4744]: E1008 09:21:10.347203 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\": container with ID starting with 00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357 not found: ID does not exist" containerID="00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.347244 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357"} err="failed to get container status \"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\": rpc error: code = NotFound desc = could not find container \"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\": container with ID starting with 00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.347272 4744 scope.go:117] "RemoveContainer" containerID="7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e" Oct 08 09:21:10 crc kubenswrapper[4744]: E1008 09:21:10.348930 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\": container with ID starting with 7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e not found: ID does not exist" containerID="7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.348958 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e"} err="failed to get container status \"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\": rpc error: code = NotFound desc = could not find container \"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\": container with ID starting with 7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.348974 4744 scope.go:117] "RemoveContainer" containerID="31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e" Oct 08 09:21:10 crc kubenswrapper[4744]: E1008 09:21:10.349282 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\": container with ID starting with 31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e not found: ID does not exist" containerID="31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.349333 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e"} err="failed to get container status \"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\": rpc error: code = NotFound desc = could not find container \"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\": container with ID starting with 31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.349413 4744 scope.go:117] "RemoveContainer" containerID="411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad" Oct 08 09:21:10 crc kubenswrapper[4744]: E1008 09:21:10.349770 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\": container with ID starting with 411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad not found: ID does not exist" containerID="411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.349810 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad"} err="failed to get container status \"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\": rpc error: code = NotFound desc = could not find container \"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\": container with ID starting with 411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.349834 4744 scope.go:117] "RemoveContainer" containerID="5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72" Oct 08 09:21:10 crc kubenswrapper[4744]: E1008 09:21:10.350117 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\": container with ID starting with 5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72 not found: ID does not exist" containerID="5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.350147 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72"} err="failed to get container status \"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\": rpc error: code = NotFound desc = could not find container \"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\": container with ID starting with 5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.350203 4744 scope.go:117] "RemoveContainer" containerID="3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb" Oct 08 09:21:10 crc kubenswrapper[4744]: E1008 09:21:10.350451 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\": container with ID starting with 3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb not found: ID does not exist" containerID="3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.350486 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb"} err="failed to get container status \"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\": rpc error: code = NotFound desc = could not find container \"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\": container with ID starting with 3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.350499 4744 scope.go:117] "RemoveContainer" containerID="0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1" Oct 08 09:21:10 crc kubenswrapper[4744]: E1008 09:21:10.351378 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\": container with ID starting with 0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1 not found: ID does not exist" containerID="0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.351408 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1"} err="failed to get container status \"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\": rpc error: code = NotFound desc = could not find container \"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\": container with ID starting with 0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.351424 4744 scope.go:117] "RemoveContainer" containerID="bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe" Oct 08 09:21:10 crc kubenswrapper[4744]: E1008 09:21:10.351688 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\": container with ID starting with bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe not found: ID does not exist" containerID="bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.351715 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe"} err="failed to get container status \"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\": rpc error: code = NotFound desc = could not find container \"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\": container with ID starting with bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.351730 4744 scope.go:117] "RemoveContainer" containerID="25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321" Oct 08 09:21:10 crc kubenswrapper[4744]: E1008 09:21:10.351965 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\": container with ID starting with 25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321 not found: ID does not exist" containerID="25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.351991 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321"} err="failed to get container status \"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\": rpc error: code = NotFound desc = could not find container \"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\": container with ID starting with 25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.352005 4744 scope.go:117] "RemoveContainer" containerID="db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.352225 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a"} err="failed to get container status \"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a\": rpc error: code = NotFound desc = could not find container \"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a\": container with ID starting with db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.352253 4744 scope.go:117] "RemoveContainer" containerID="00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.352481 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357"} err="failed to get container status \"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\": rpc error: code = NotFound desc = could not find container \"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\": container with ID starting with 00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.352501 4744 scope.go:117] "RemoveContainer" containerID="7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.352689 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e"} err="failed to get container status \"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\": rpc error: code = NotFound desc = could not find container \"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\": container with ID starting with 7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.352706 4744 scope.go:117] "RemoveContainer" containerID="31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.352867 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e"} err="failed to get container status \"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\": rpc error: code = NotFound desc = could not find container \"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\": container with ID starting with 31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.352888 4744 scope.go:117] "RemoveContainer" containerID="411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.353052 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad"} err="failed to get container status \"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\": rpc error: code = NotFound desc = could not find container \"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\": container with ID starting with 411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.353070 4744 scope.go:117] "RemoveContainer" containerID="5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.353228 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72"} err="failed to get container status \"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\": rpc error: code = NotFound desc = could not find container \"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\": container with ID starting with 5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.353244 4744 scope.go:117] "RemoveContainer" containerID="3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.353434 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb"} err="failed to get container status \"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\": rpc error: code = NotFound desc = could not find container \"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\": container with ID starting with 3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.353451 4744 scope.go:117] "RemoveContainer" containerID="0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.353660 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1"} err="failed to get container status \"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\": rpc error: code = NotFound desc = could not find container \"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\": container with ID starting with 0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.353685 4744 scope.go:117] "RemoveContainer" containerID="bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.353899 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe"} err="failed to get container status \"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\": rpc error: code = NotFound desc = could not find container \"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\": container with ID starting with bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.353918 4744 scope.go:117] "RemoveContainer" containerID="25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.354133 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321"} err="failed to get container status \"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\": rpc error: code = NotFound desc = could not find container \"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\": container with ID starting with 25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.354171 4744 scope.go:117] "RemoveContainer" containerID="db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.354388 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a"} err="failed to get container status \"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a\": rpc error: code = NotFound desc = could not find container \"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a\": container with ID starting with db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.354413 4744 scope.go:117] "RemoveContainer" containerID="00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.354609 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357"} err="failed to get container status \"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\": rpc error: code = NotFound desc = could not find container \"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\": container with ID starting with 00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.354628 4744 scope.go:117] "RemoveContainer" containerID="7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.354820 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e"} err="failed to get container status \"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\": rpc error: code = NotFound desc = could not find container \"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\": container with ID starting with 7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.354839 4744 scope.go:117] "RemoveContainer" containerID="31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.355016 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e"} err="failed to get container status \"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\": rpc error: code = NotFound desc = could not find container \"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\": container with ID starting with 31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.355038 4744 scope.go:117] "RemoveContainer" containerID="411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.355254 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad"} err="failed to get container status \"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\": rpc error: code = NotFound desc = could not find container \"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\": container with ID starting with 411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.355275 4744 scope.go:117] "RemoveContainer" containerID="5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.355473 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72"} err="failed to get container status \"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\": rpc error: code = NotFound desc = could not find container \"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\": container with ID starting with 5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.355492 4744 scope.go:117] "RemoveContainer" containerID="3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.355710 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb"} err="failed to get container status \"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\": rpc error: code = NotFound desc = could not find container \"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\": container with ID starting with 3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.355733 4744 scope.go:117] "RemoveContainer" containerID="0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.357472 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1"} err="failed to get container status \"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\": rpc error: code = NotFound desc = could not find container \"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\": container with ID starting with 0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.357493 4744 scope.go:117] "RemoveContainer" containerID="bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.357947 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe"} err="failed to get container status \"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\": rpc error: code = NotFound desc = could not find container \"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\": container with ID starting with bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.357974 4744 scope.go:117] "RemoveContainer" containerID="25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.358673 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321"} err="failed to get container status \"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\": rpc error: code = NotFound desc = could not find container \"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\": container with ID starting with 25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.358692 4744 scope.go:117] "RemoveContainer" containerID="db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.359006 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a"} err="failed to get container status \"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a\": rpc error: code = NotFound desc = could not find container \"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a\": container with ID starting with db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.359029 4744 scope.go:117] "RemoveContainer" containerID="00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.359350 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357"} err="failed to get container status \"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\": rpc error: code = NotFound desc = could not find container \"00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357\": container with ID starting with 00d5600ff81698182a15202c736309226f05ee24b3a7eed8c915cbfa51dbf357 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.359437 4744 scope.go:117] "RemoveContainer" containerID="7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.359755 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e"} err="failed to get container status \"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\": rpc error: code = NotFound desc = could not find container \"7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e\": container with ID starting with 7db290b150a770c51bb06b2cafa9daf211d23b304513d9ecb63d918902f1286e not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.359776 4744 scope.go:117] "RemoveContainer" containerID="31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.362001 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e"} err="failed to get container status \"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\": rpc error: code = NotFound desc = could not find container \"31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e\": container with ID starting with 31584ca3993974b9a4a8fc816e14caa5cd6e2c16e524f6ee6ae4a3e2f8592a6e not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.362029 4744 scope.go:117] "RemoveContainer" containerID="411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.362513 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad"} err="failed to get container status \"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\": rpc error: code = NotFound desc = could not find container \"411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad\": container with ID starting with 411ba2bba159d1b42e367136e77fd98d8f9ce7ac7c68c0eb0024fa3f515024ad not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.362567 4744 scope.go:117] "RemoveContainer" containerID="5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.362921 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72"} err="failed to get container status \"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\": rpc error: code = NotFound desc = could not find container \"5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72\": container with ID starting with 5d2dba222807005de8c3418ef3567c30a1c820e930be8043241beff0f47d1f72 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.362965 4744 scope.go:117] "RemoveContainer" containerID="3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.363307 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb"} err="failed to get container status \"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\": rpc error: code = NotFound desc = could not find container \"3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb\": container with ID starting with 3e3aa8469d1c0a6a2a1278bca42aee7d18ef9696d7fb21cb4f9549f84cef0ebb not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.363329 4744 scope.go:117] "RemoveContainer" containerID="0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.363619 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1"} err="failed to get container status \"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\": rpc error: code = NotFound desc = could not find container \"0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1\": container with ID starting with 0d44a26fee516bf471ed8d305a539ec26fc102235f35948f95b7d082cf8de0f1 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.363642 4744 scope.go:117] "RemoveContainer" containerID="bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.363881 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe"} err="failed to get container status \"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\": rpc error: code = NotFound desc = could not find container \"bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe\": container with ID starting with bfb436ef0a2b0536d1131f209978cb1b1f4e2e35756ae8666190ca9de5ee8ffe not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.363907 4744 scope.go:117] "RemoveContainer" containerID="25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.364125 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321"} err="failed to get container status \"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\": rpc error: code = NotFound desc = could not find container \"25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321\": container with ID starting with 25490e2698d2a29e7210cf00382342b3736b1eebfb16effb704be3990f5f2321 not found: ID does not exist" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.364142 4744 scope.go:117] "RemoveContainer" containerID="db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a" Oct 08 09:21:10 crc kubenswrapper[4744]: I1008 09:21:10.364299 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a"} err="failed to get container status \"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a\": rpc error: code = NotFound desc = could not find container \"db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a\": container with ID starting with db84fc0cced7a4d7e9d70a4e55f02377b34fce7a8c0a02acc865880a5185aa6a not found: ID does not exist" Oct 08 09:21:11 crc kubenswrapper[4744]: I1008 09:21:11.049332 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" event={"ID":"1eb3be1b-60dd-44d8-b727-1cab861c60e5","Type":"ContainerStarted","Data":"84d5c4b464000d168817b6a679801585601d07e219f94a2eedbc54c9e4bca76d"} Oct 08 09:21:11 crc kubenswrapper[4744]: I1008 09:21:11.049895 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" event={"ID":"1eb3be1b-60dd-44d8-b727-1cab861c60e5","Type":"ContainerStarted","Data":"ecd8415a57e619c7417204b32cfa552be87fc43c614f5907fa11639d4f13204c"} Oct 08 09:21:11 crc kubenswrapper[4744]: I1008 09:21:11.049920 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" event={"ID":"1eb3be1b-60dd-44d8-b727-1cab861c60e5","Type":"ContainerStarted","Data":"d75ca8b1a4d2a43c203951bfa4e811c540ca2fecbcec3b39444a650a21b9c094"} Oct 08 09:21:11 crc kubenswrapper[4744]: I1008 09:21:11.049940 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" event={"ID":"1eb3be1b-60dd-44d8-b727-1cab861c60e5","Type":"ContainerStarted","Data":"3bda8bab5f859dda901e6f0954cfd66901fa0f9f4e9e84f87ef7045a9b880a39"} Oct 08 09:21:11 crc kubenswrapper[4744]: I1008 09:21:11.049958 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" event={"ID":"1eb3be1b-60dd-44d8-b727-1cab861c60e5","Type":"ContainerStarted","Data":"046699ffebd1f441f6ebc209a1b407fe5cc8c39dbfc5da4571049120b2ac6e5e"} Oct 08 09:21:11 crc kubenswrapper[4744]: I1008 09:21:11.049981 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" event={"ID":"1eb3be1b-60dd-44d8-b727-1cab861c60e5","Type":"ContainerStarted","Data":"1005cf632ba33a17abed070adb09cbe9bd7794e896b4d791576477906711535c"} Oct 08 09:21:11 crc kubenswrapper[4744]: I1008 09:21:11.462994 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a9bf215-f18b-4732-b12c-0355a3a0a15f" path="/var/lib/kubelet/pods/5a9bf215-f18b-4732-b12c-0355a3a0a15f/volumes" Oct 08 09:21:14 crc kubenswrapper[4744]: I1008 09:21:14.091147 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" event={"ID":"1eb3be1b-60dd-44d8-b727-1cab861c60e5","Type":"ContainerStarted","Data":"2830f4af73146880555d25e379ce136c5caabadccef5c314a0075b32690853bc"} Oct 08 09:21:16 crc kubenswrapper[4744]: I1008 09:21:16.105699 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" event={"ID":"1eb3be1b-60dd-44d8-b727-1cab861c60e5","Type":"ContainerStarted","Data":"f1447fb6989262af2e727e0747affaefb29f6586b4e97cb0458abc48fe222434"} Oct 08 09:21:16 crc kubenswrapper[4744]: I1008 09:21:16.106611 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:16 crc kubenswrapper[4744]: I1008 09:21:16.106632 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:16 crc kubenswrapper[4744]: I1008 09:21:16.106643 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:16 crc kubenswrapper[4744]: I1008 09:21:16.134213 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:16 crc kubenswrapper[4744]: I1008 09:21:16.134527 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:16 crc kubenswrapper[4744]: I1008 09:21:16.144336 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" podStartSLOduration=7.144317921 podStartE2EDuration="7.144317921s" podCreationTimestamp="2025-10-08 09:21:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:21:16.142289032 +0000 UTC m=+571.389934291" watchObservedRunningTime="2025-10-08 09:21:16.144317921 +0000 UTC m=+571.391963160" Oct 08 09:21:19 crc kubenswrapper[4744]: I1008 09:21:19.691091 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:21:19 crc kubenswrapper[4744]: I1008 09:21:19.691212 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:21:19 crc kubenswrapper[4744]: I1008 09:21:19.691295 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:21:19 crc kubenswrapper[4744]: I1008 09:21:19.693039 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c068a91bb7d0e26249b43fab5fe60ecb8e806b9d40583c0517d3afb38ddf715"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 09:21:19 crc kubenswrapper[4744]: I1008 09:21:19.693162 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://8c068a91bb7d0e26249b43fab5fe60ecb8e806b9d40583c0517d3afb38ddf715" gracePeriod=600 Oct 08 09:21:20 crc kubenswrapper[4744]: I1008 09:21:20.135350 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="8c068a91bb7d0e26249b43fab5fe60ecb8e806b9d40583c0517d3afb38ddf715" exitCode=0 Oct 08 09:21:20 crc kubenswrapper[4744]: I1008 09:21:20.135489 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"8c068a91bb7d0e26249b43fab5fe60ecb8e806b9d40583c0517d3afb38ddf715"} Oct 08 09:21:20 crc kubenswrapper[4744]: I1008 09:21:20.135709 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"12c8d8cd4abdd9a59c0cfbc9852b990e3820f44c3e436a88bb84a00c67314d34"} Oct 08 09:21:20 crc kubenswrapper[4744]: I1008 09:21:20.135736 4744 scope.go:117] "RemoveContainer" containerID="63e1dfa02860e1e750f72656194fd55b1911a02f76a9cd7c271f9f20e1dffa0a" Oct 08 09:21:25 crc kubenswrapper[4744]: I1008 09:21:25.459245 4744 scope.go:117] "RemoveContainer" containerID="94b68c9a4976717260751eead14fea87839d1758f32a92629bc6776f522afb3e" Oct 08 09:21:25 crc kubenswrapper[4744]: E1008 09:21:25.465647 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-dldlq_openshift-multus(7ea48507-1f33-46fc-ab34-c6e083dd86a9)\"" pod="openshift-multus/multus-dldlq" podUID="7ea48507-1f33-46fc-ab34-c6e083dd86a9" Oct 08 09:21:36 crc kubenswrapper[4744]: I1008 09:21:36.452961 4744 scope.go:117] "RemoveContainer" containerID="94b68c9a4976717260751eead14fea87839d1758f32a92629bc6776f522afb3e" Oct 08 09:21:37 crc kubenswrapper[4744]: I1008 09:21:37.261492 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dldlq_7ea48507-1f33-46fc-ab34-c6e083dd86a9/kube-multus/2.log" Oct 08 09:21:37 crc kubenswrapper[4744]: I1008 09:21:37.262486 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dldlq_7ea48507-1f33-46fc-ab34-c6e083dd86a9/kube-multus/1.log" Oct 08 09:21:37 crc kubenswrapper[4744]: I1008 09:21:37.262562 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dldlq" event={"ID":"7ea48507-1f33-46fc-ab34-c6e083dd86a9","Type":"ContainerStarted","Data":"ebb0e05ad9930aca477695be6d3ced04726ea9fdfcb775afc5db8c08307524a5"} Oct 08 09:21:39 crc kubenswrapper[4744]: I1008 09:21:39.768037 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cjbt2" Oct 08 09:21:45 crc kubenswrapper[4744]: I1008 09:21:45.644820 4744 scope.go:117] "RemoveContainer" containerID="6387ad8ee2c14d3df619d9dd713e9f263a5e06162f8937a8ecbdd1a65322ea75" Oct 08 09:21:46 crc kubenswrapper[4744]: I1008 09:21:46.317646 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dldlq_7ea48507-1f33-46fc-ab34-c6e083dd86a9/kube-multus/2.log" Oct 08 09:21:49 crc kubenswrapper[4744]: I1008 09:21:49.773957 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65"] Oct 08 09:21:49 crc kubenswrapper[4744]: I1008 09:21:49.777834 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:49 crc kubenswrapper[4744]: I1008 09:21:49.786421 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65"] Oct 08 09:21:49 crc kubenswrapper[4744]: I1008 09:21:49.787725 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 09:21:49 crc kubenswrapper[4744]: I1008 09:21:49.894143 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8a3e5f3c-8790-47bb-a646-97dbbf595559-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65\" (UID: \"8a3e5f3c-8790-47bb-a646-97dbbf595559\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:49 crc kubenswrapper[4744]: I1008 09:21:49.894460 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwp54\" (UniqueName: \"kubernetes.io/projected/8a3e5f3c-8790-47bb-a646-97dbbf595559-kube-api-access-hwp54\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65\" (UID: \"8a3e5f3c-8790-47bb-a646-97dbbf595559\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:49 crc kubenswrapper[4744]: I1008 09:21:49.894486 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8a3e5f3c-8790-47bb-a646-97dbbf595559-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65\" (UID: \"8a3e5f3c-8790-47bb-a646-97dbbf595559\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:49 crc kubenswrapper[4744]: I1008 09:21:49.995442 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwp54\" (UniqueName: \"kubernetes.io/projected/8a3e5f3c-8790-47bb-a646-97dbbf595559-kube-api-access-hwp54\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65\" (UID: \"8a3e5f3c-8790-47bb-a646-97dbbf595559\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:49 crc kubenswrapper[4744]: I1008 09:21:49.995544 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8a3e5f3c-8790-47bb-a646-97dbbf595559-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65\" (UID: \"8a3e5f3c-8790-47bb-a646-97dbbf595559\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:49 crc kubenswrapper[4744]: I1008 09:21:49.995600 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8a3e5f3c-8790-47bb-a646-97dbbf595559-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65\" (UID: \"8a3e5f3c-8790-47bb-a646-97dbbf595559\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:49 crc kubenswrapper[4744]: I1008 09:21:49.996349 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8a3e5f3c-8790-47bb-a646-97dbbf595559-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65\" (UID: \"8a3e5f3c-8790-47bb-a646-97dbbf595559\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:49 crc kubenswrapper[4744]: I1008 09:21:49.997248 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8a3e5f3c-8790-47bb-a646-97dbbf595559-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65\" (UID: \"8a3e5f3c-8790-47bb-a646-97dbbf595559\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:50 crc kubenswrapper[4744]: I1008 09:21:50.016767 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwp54\" (UniqueName: \"kubernetes.io/projected/8a3e5f3c-8790-47bb-a646-97dbbf595559-kube-api-access-hwp54\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65\" (UID: \"8a3e5f3c-8790-47bb-a646-97dbbf595559\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:50 crc kubenswrapper[4744]: I1008 09:21:50.096267 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:50 crc kubenswrapper[4744]: I1008 09:21:50.415169 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65"] Oct 08 09:21:51 crc kubenswrapper[4744]: I1008 09:21:51.347474 4744 generic.go:334] "Generic (PLEG): container finished" podID="8a3e5f3c-8790-47bb-a646-97dbbf595559" containerID="2c0cd7974c35f0110042136f983361154a94410ebdbd27c07f923f3e94c93521" exitCode=0 Oct 08 09:21:51 crc kubenswrapper[4744]: I1008 09:21:51.347526 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" event={"ID":"8a3e5f3c-8790-47bb-a646-97dbbf595559","Type":"ContainerDied","Data":"2c0cd7974c35f0110042136f983361154a94410ebdbd27c07f923f3e94c93521"} Oct 08 09:21:51 crc kubenswrapper[4744]: I1008 09:21:51.347554 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" event={"ID":"8a3e5f3c-8790-47bb-a646-97dbbf595559","Type":"ContainerStarted","Data":"d5b07c247ded398e6d77bbb147aa66108c1daab8e074128047cd8fadaf572642"} Oct 08 09:21:53 crc kubenswrapper[4744]: I1008 09:21:53.364233 4744 generic.go:334] "Generic (PLEG): container finished" podID="8a3e5f3c-8790-47bb-a646-97dbbf595559" containerID="7cf8492fc6a037353e76204646d93a7f8a4ec8dc4ee4636a2da99fb49f18ddf4" exitCode=0 Oct 08 09:21:53 crc kubenswrapper[4744]: I1008 09:21:53.364353 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" event={"ID":"8a3e5f3c-8790-47bb-a646-97dbbf595559","Type":"ContainerDied","Data":"7cf8492fc6a037353e76204646d93a7f8a4ec8dc4ee4636a2da99fb49f18ddf4"} Oct 08 09:21:54 crc kubenswrapper[4744]: I1008 09:21:54.373355 4744 generic.go:334] "Generic (PLEG): container finished" podID="8a3e5f3c-8790-47bb-a646-97dbbf595559" containerID="d1819731dca3e2ec26ce2fde0662e1cde1e6dbf8ea1c9f477ffdd2c9b09dcc22" exitCode=0 Oct 08 09:21:54 crc kubenswrapper[4744]: I1008 09:21:54.373500 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" event={"ID":"8a3e5f3c-8790-47bb-a646-97dbbf595559","Type":"ContainerDied","Data":"d1819731dca3e2ec26ce2fde0662e1cde1e6dbf8ea1c9f477ffdd2c9b09dcc22"} Oct 08 09:21:55 crc kubenswrapper[4744]: I1008 09:21:55.596341 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:55 crc kubenswrapper[4744]: I1008 09:21:55.776028 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8a3e5f3c-8790-47bb-a646-97dbbf595559-util\") pod \"8a3e5f3c-8790-47bb-a646-97dbbf595559\" (UID: \"8a3e5f3c-8790-47bb-a646-97dbbf595559\") " Oct 08 09:21:55 crc kubenswrapper[4744]: I1008 09:21:55.776147 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8a3e5f3c-8790-47bb-a646-97dbbf595559-bundle\") pod \"8a3e5f3c-8790-47bb-a646-97dbbf595559\" (UID: \"8a3e5f3c-8790-47bb-a646-97dbbf595559\") " Oct 08 09:21:55 crc kubenswrapper[4744]: I1008 09:21:55.776189 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwp54\" (UniqueName: \"kubernetes.io/projected/8a3e5f3c-8790-47bb-a646-97dbbf595559-kube-api-access-hwp54\") pod \"8a3e5f3c-8790-47bb-a646-97dbbf595559\" (UID: \"8a3e5f3c-8790-47bb-a646-97dbbf595559\") " Oct 08 09:21:55 crc kubenswrapper[4744]: I1008 09:21:55.777326 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a3e5f3c-8790-47bb-a646-97dbbf595559-bundle" (OuterVolumeSpecName: "bundle") pod "8a3e5f3c-8790-47bb-a646-97dbbf595559" (UID: "8a3e5f3c-8790-47bb-a646-97dbbf595559"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:21:55 crc kubenswrapper[4744]: I1008 09:21:55.785766 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a3e5f3c-8790-47bb-a646-97dbbf595559-kube-api-access-hwp54" (OuterVolumeSpecName: "kube-api-access-hwp54") pod "8a3e5f3c-8790-47bb-a646-97dbbf595559" (UID: "8a3e5f3c-8790-47bb-a646-97dbbf595559"). InnerVolumeSpecName "kube-api-access-hwp54". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:21:55 crc kubenswrapper[4744]: I1008 09:21:55.877822 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8a3e5f3c-8790-47bb-a646-97dbbf595559-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:55 crc kubenswrapper[4744]: I1008 09:21:55.877862 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwp54\" (UniqueName: \"kubernetes.io/projected/8a3e5f3c-8790-47bb-a646-97dbbf595559-kube-api-access-hwp54\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:56 crc kubenswrapper[4744]: I1008 09:21:56.059484 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a3e5f3c-8790-47bb-a646-97dbbf595559-util" (OuterVolumeSpecName: "util") pod "8a3e5f3c-8790-47bb-a646-97dbbf595559" (UID: "8a3e5f3c-8790-47bb-a646-97dbbf595559"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:21:56 crc kubenswrapper[4744]: I1008 09:21:56.080699 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8a3e5f3c-8790-47bb-a646-97dbbf595559-util\") on node \"crc\" DevicePath \"\"" Oct 08 09:21:56 crc kubenswrapper[4744]: I1008 09:21:56.395194 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" event={"ID":"8a3e5f3c-8790-47bb-a646-97dbbf595559","Type":"ContainerDied","Data":"d5b07c247ded398e6d77bbb147aa66108c1daab8e074128047cd8fadaf572642"} Oct 08 09:21:56 crc kubenswrapper[4744]: I1008 09:21:56.395489 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5b07c247ded398e6d77bbb147aa66108c1daab8e074128047cd8fadaf572642" Oct 08 09:21:56 crc kubenswrapper[4744]: I1008 09:21:56.395246 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.426819 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-nmnhx"] Oct 08 09:21:57 crc kubenswrapper[4744]: E1008 09:21:57.427071 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a3e5f3c-8790-47bb-a646-97dbbf595559" containerName="extract" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.427086 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a3e5f3c-8790-47bb-a646-97dbbf595559" containerName="extract" Oct 08 09:21:57 crc kubenswrapper[4744]: E1008 09:21:57.427099 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a3e5f3c-8790-47bb-a646-97dbbf595559" containerName="pull" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.427108 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a3e5f3c-8790-47bb-a646-97dbbf595559" containerName="pull" Oct 08 09:21:57 crc kubenswrapper[4744]: E1008 09:21:57.427122 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a3e5f3c-8790-47bb-a646-97dbbf595559" containerName="util" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.427130 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a3e5f3c-8790-47bb-a646-97dbbf595559" containerName="util" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.427255 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a3e5f3c-8790-47bb-a646-97dbbf595559" containerName="extract" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.427738 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-nmnhx" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.429847 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-q6nk7" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.430276 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.431643 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.444593 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-nmnhx"] Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.499332 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jklp7\" (UniqueName: \"kubernetes.io/projected/1f1d4ca0-878e-4211-a40a-37f16085fe47-kube-api-access-jklp7\") pod \"nmstate-operator-858ddd8f98-nmnhx\" (UID: \"1f1d4ca0-878e-4211-a40a-37f16085fe47\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-nmnhx" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.600538 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jklp7\" (UniqueName: \"kubernetes.io/projected/1f1d4ca0-878e-4211-a40a-37f16085fe47-kube-api-access-jklp7\") pod \"nmstate-operator-858ddd8f98-nmnhx\" (UID: \"1f1d4ca0-878e-4211-a40a-37f16085fe47\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-nmnhx" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.625713 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jklp7\" (UniqueName: \"kubernetes.io/projected/1f1d4ca0-878e-4211-a40a-37f16085fe47-kube-api-access-jklp7\") pod \"nmstate-operator-858ddd8f98-nmnhx\" (UID: \"1f1d4ca0-878e-4211-a40a-37f16085fe47\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-nmnhx" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.743649 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-nmnhx" Oct 08 09:21:57 crc kubenswrapper[4744]: I1008 09:21:57.929396 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-nmnhx"] Oct 08 09:21:58 crc kubenswrapper[4744]: I1008 09:21:58.406868 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-nmnhx" event={"ID":"1f1d4ca0-878e-4211-a40a-37f16085fe47","Type":"ContainerStarted","Data":"b3d49070724a5b093a08dbae791f2ed112fca22399cbaeacbc105146a05c9f48"} Oct 08 09:22:01 crc kubenswrapper[4744]: I1008 09:22:01.434430 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-nmnhx" event={"ID":"1f1d4ca0-878e-4211-a40a-37f16085fe47","Type":"ContainerStarted","Data":"1d74d5db1bcc479a4ae395a84bab3dfb17b64d9141ac83a676539741f110a7de"} Oct 08 09:22:01 crc kubenswrapper[4744]: I1008 09:22:01.456960 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-nmnhx" podStartSLOduration=2.142869217 podStartE2EDuration="4.45693774s" podCreationTimestamp="2025-10-08 09:21:57 +0000 UTC" firstStartedPulling="2025-10-08 09:21:57.949047485 +0000 UTC m=+613.196692724" lastFinishedPulling="2025-10-08 09:22:00.263116008 +0000 UTC m=+615.510761247" observedRunningTime="2025-10-08 09:22:01.449499735 +0000 UTC m=+616.697144984" watchObservedRunningTime="2025-10-08 09:22:01.45693774 +0000 UTC m=+616.704582989" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.425227 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-9qfcx"] Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.426318 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-9qfcx" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.430747 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-nhw7t" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.441258 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l"] Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.442204 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.447891 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-9qfcx"] Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.449024 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.467098 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29lqd\" (UniqueName: \"kubernetes.io/projected/89b615a3-0a5d-483c-8d41-2dee267f1686-kube-api-access-29lqd\") pod \"nmstate-webhook-6cdbc54649-pnz4l\" (UID: \"89b615a3-0a5d-483c-8d41-2dee267f1686\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.467170 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42hqn\" (UniqueName: \"kubernetes.io/projected/db58dfb9-4fd9-4896-add0-116a1ed9079d-kube-api-access-42hqn\") pod \"nmstate-metrics-fdff9cb8d-9qfcx\" (UID: \"db58dfb9-4fd9-4896-add0-116a1ed9079d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-9qfcx" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.467208 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/89b615a3-0a5d-483c-8d41-2dee267f1686-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-pnz4l\" (UID: \"89b615a3-0a5d-483c-8d41-2dee267f1686\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.476998 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l"] Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.498428 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-tk2k5"] Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.499274 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.568719 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/89b615a3-0a5d-483c-8d41-2dee267f1686-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-pnz4l\" (UID: \"89b615a3-0a5d-483c-8d41-2dee267f1686\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.568775 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8b5fc549-4718-4e89-82a9-b39e93a2220d-ovs-socket\") pod \"nmstate-handler-tk2k5\" (UID: \"8b5fc549-4718-4e89-82a9-b39e93a2220d\") " pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.568803 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8b5fc549-4718-4e89-82a9-b39e93a2220d-nmstate-lock\") pod \"nmstate-handler-tk2k5\" (UID: \"8b5fc549-4718-4e89-82a9-b39e93a2220d\") " pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.568838 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spvqw\" (UniqueName: \"kubernetes.io/projected/8b5fc549-4718-4e89-82a9-b39e93a2220d-kube-api-access-spvqw\") pod \"nmstate-handler-tk2k5\" (UID: \"8b5fc549-4718-4e89-82a9-b39e93a2220d\") " pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.568873 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29lqd\" (UniqueName: \"kubernetes.io/projected/89b615a3-0a5d-483c-8d41-2dee267f1686-kube-api-access-29lqd\") pod \"nmstate-webhook-6cdbc54649-pnz4l\" (UID: \"89b615a3-0a5d-483c-8d41-2dee267f1686\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.568898 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8b5fc549-4718-4e89-82a9-b39e93a2220d-dbus-socket\") pod \"nmstate-handler-tk2k5\" (UID: \"8b5fc549-4718-4e89-82a9-b39e93a2220d\") " pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.568918 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42hqn\" (UniqueName: \"kubernetes.io/projected/db58dfb9-4fd9-4896-add0-116a1ed9079d-kube-api-access-42hqn\") pod \"nmstate-metrics-fdff9cb8d-9qfcx\" (UID: \"db58dfb9-4fd9-4896-add0-116a1ed9079d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-9qfcx" Oct 08 09:22:02 crc kubenswrapper[4744]: E1008 09:22:02.569236 4744 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 08 09:22:02 crc kubenswrapper[4744]: E1008 09:22:02.569283 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89b615a3-0a5d-483c-8d41-2dee267f1686-tls-key-pair podName:89b615a3-0a5d-483c-8d41-2dee267f1686 nodeName:}" failed. No retries permitted until 2025-10-08 09:22:03.069267932 +0000 UTC m=+618.316913171 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/89b615a3-0a5d-483c-8d41-2dee267f1686-tls-key-pair") pod "nmstate-webhook-6cdbc54649-pnz4l" (UID: "89b615a3-0a5d-483c-8d41-2dee267f1686") : secret "openshift-nmstate-webhook" not found Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.590168 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29lqd\" (UniqueName: \"kubernetes.io/projected/89b615a3-0a5d-483c-8d41-2dee267f1686-kube-api-access-29lqd\") pod \"nmstate-webhook-6cdbc54649-pnz4l\" (UID: \"89b615a3-0a5d-483c-8d41-2dee267f1686\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.590902 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42hqn\" (UniqueName: \"kubernetes.io/projected/db58dfb9-4fd9-4896-add0-116a1ed9079d-kube-api-access-42hqn\") pod \"nmstate-metrics-fdff9cb8d-9qfcx\" (UID: \"db58dfb9-4fd9-4896-add0-116a1ed9079d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-9qfcx" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.646299 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff"] Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.647014 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.653650 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.657108 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.657431 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-cb6tk" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.664541 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff"] Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.670271 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8b5fc549-4718-4e89-82a9-b39e93a2220d-dbus-socket\") pod \"nmstate-handler-tk2k5\" (UID: \"8b5fc549-4718-4e89-82a9-b39e93a2220d\") " pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.670543 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8b5fc549-4718-4e89-82a9-b39e93a2220d-ovs-socket\") pod \"nmstate-handler-tk2k5\" (UID: \"8b5fc549-4718-4e89-82a9-b39e93a2220d\") " pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.670637 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8b5fc549-4718-4e89-82a9-b39e93a2220d-nmstate-lock\") pod \"nmstate-handler-tk2k5\" (UID: \"8b5fc549-4718-4e89-82a9-b39e93a2220d\") " pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.670666 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8b5fc549-4718-4e89-82a9-b39e93a2220d-dbus-socket\") pod \"nmstate-handler-tk2k5\" (UID: \"8b5fc549-4718-4e89-82a9-b39e93a2220d\") " pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.670699 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8b5fc549-4718-4e89-82a9-b39e93a2220d-ovs-socket\") pod \"nmstate-handler-tk2k5\" (UID: \"8b5fc549-4718-4e89-82a9-b39e93a2220d\") " pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.670738 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8b5fc549-4718-4e89-82a9-b39e93a2220d-nmstate-lock\") pod \"nmstate-handler-tk2k5\" (UID: \"8b5fc549-4718-4e89-82a9-b39e93a2220d\") " pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.670794 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spvqw\" (UniqueName: \"kubernetes.io/projected/8b5fc549-4718-4e89-82a9-b39e93a2220d-kube-api-access-spvqw\") pod \"nmstate-handler-tk2k5\" (UID: \"8b5fc549-4718-4e89-82a9-b39e93a2220d\") " pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.703044 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spvqw\" (UniqueName: \"kubernetes.io/projected/8b5fc549-4718-4e89-82a9-b39e93a2220d-kube-api-access-spvqw\") pod \"nmstate-handler-tk2k5\" (UID: \"8b5fc549-4718-4e89-82a9-b39e93a2220d\") " pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.740365 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-9qfcx" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.773357 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c9804b5-0729-4059-bb00-35fd1d355b3c-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-hrjff\" (UID: \"2c9804b5-0729-4059-bb00-35fd1d355b3c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.773491 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nd87\" (UniqueName: \"kubernetes.io/projected/2c9804b5-0729-4059-bb00-35fd1d355b3c-kube-api-access-6nd87\") pod \"nmstate-console-plugin-6b874cbd85-hrjff\" (UID: \"2c9804b5-0729-4059-bb00-35fd1d355b3c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.773524 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2c9804b5-0729-4059-bb00-35fd1d355b3c-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-hrjff\" (UID: \"2c9804b5-0729-4059-bb00-35fd1d355b3c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.812179 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.874215 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c9804b5-0729-4059-bb00-35fd1d355b3c-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-hrjff\" (UID: \"2c9804b5-0729-4059-bb00-35fd1d355b3c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.874307 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nd87\" (UniqueName: \"kubernetes.io/projected/2c9804b5-0729-4059-bb00-35fd1d355b3c-kube-api-access-6nd87\") pod \"nmstate-console-plugin-6b874cbd85-hrjff\" (UID: \"2c9804b5-0729-4059-bb00-35fd1d355b3c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.874329 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2c9804b5-0729-4059-bb00-35fd1d355b3c-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-hrjff\" (UID: \"2c9804b5-0729-4059-bb00-35fd1d355b3c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.875591 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/2c9804b5-0729-4059-bb00-35fd1d355b3c-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-hrjff\" (UID: \"2c9804b5-0729-4059-bb00-35fd1d355b3c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.880608 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c9804b5-0729-4059-bb00-35fd1d355b3c-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-hrjff\" (UID: \"2c9804b5-0729-4059-bb00-35fd1d355b3c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.896910 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nd87\" (UniqueName: \"kubernetes.io/projected/2c9804b5-0729-4059-bb00-35fd1d355b3c-kube-api-access-6nd87\") pod \"nmstate-console-plugin-6b874cbd85-hrjff\" (UID: \"2c9804b5-0729-4059-bb00-35fd1d355b3c\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.922828 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7cbf967b4c-hpjvs"] Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.923715 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.945193 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7cbf967b4c-hpjvs"] Oct 08 09:22:02 crc kubenswrapper[4744]: I1008 09:22:02.963468 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.069747 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-9qfcx"] Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.104818 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb9f4b9c-410e-4b88-8172-f65d81b767f3-service-ca\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.104893 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb9f4b9c-410e-4b88-8172-f65d81b767f3-trusted-ca-bundle\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.104918 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb9f4b9c-410e-4b88-8172-f65d81b767f3-console-config\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.104948 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb9f4b9c-410e-4b88-8172-f65d81b767f3-console-oauth-config\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.104978 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb9f4b9c-410e-4b88-8172-f65d81b767f3-oauth-serving-cert\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.104997 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mjjj\" (UniqueName: \"kubernetes.io/projected/bb9f4b9c-410e-4b88-8172-f65d81b767f3-kube-api-access-7mjjj\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.105025 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/89b615a3-0a5d-483c-8d41-2dee267f1686-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-pnz4l\" (UID: \"89b615a3-0a5d-483c-8d41-2dee267f1686\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.105049 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb9f4b9c-410e-4b88-8172-f65d81b767f3-console-serving-cert\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.114584 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/89b615a3-0a5d-483c-8d41-2dee267f1686-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-pnz4l\" (UID: \"89b615a3-0a5d-483c-8d41-2dee267f1686\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.205794 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb9f4b9c-410e-4b88-8172-f65d81b767f3-trusted-ca-bundle\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.205839 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb9f4b9c-410e-4b88-8172-f65d81b767f3-console-config\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.205875 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb9f4b9c-410e-4b88-8172-f65d81b767f3-console-oauth-config\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.205906 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb9f4b9c-410e-4b88-8172-f65d81b767f3-oauth-serving-cert\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.205928 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mjjj\" (UniqueName: \"kubernetes.io/projected/bb9f4b9c-410e-4b88-8172-f65d81b767f3-kube-api-access-7mjjj\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.205951 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb9f4b9c-410e-4b88-8172-f65d81b767f3-console-serving-cert\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.205976 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb9f4b9c-410e-4b88-8172-f65d81b767f3-service-ca\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.207037 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb9f4b9c-410e-4b88-8172-f65d81b767f3-trusted-ca-bundle\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.207142 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb9f4b9c-410e-4b88-8172-f65d81b767f3-service-ca\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.207340 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb9f4b9c-410e-4b88-8172-f65d81b767f3-console-config\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.207846 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb9f4b9c-410e-4b88-8172-f65d81b767f3-oauth-serving-cert\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.210973 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb9f4b9c-410e-4b88-8172-f65d81b767f3-console-oauth-config\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.211296 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb9f4b9c-410e-4b88-8172-f65d81b767f3-console-serving-cert\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.230336 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mjjj\" (UniqueName: \"kubernetes.io/projected/bb9f4b9c-410e-4b88-8172-f65d81b767f3-kube-api-access-7mjjj\") pod \"console-7cbf967b4c-hpjvs\" (UID: \"bb9f4b9c-410e-4b88-8172-f65d81b767f3\") " pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.234154 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff"] Oct 08 09:22:03 crc kubenswrapper[4744]: W1008 09:22:03.238665 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c9804b5_0729_4059_bb00_35fd1d355b3c.slice/crio-6ceefce5bc10f1c95777ae3f15fa0ced244c95d5367e45e7732d0eb1a11d9757 WatchSource:0}: Error finding container 6ceefce5bc10f1c95777ae3f15fa0ced244c95d5367e45e7732d0eb1a11d9757: Status 404 returned error can't find the container with id 6ceefce5bc10f1c95777ae3f15fa0ced244c95d5367e45e7732d0eb1a11d9757 Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.253326 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.357985 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.448971 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7cbf967b4c-hpjvs"] Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.478326 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-tk2k5" event={"ID":"8b5fc549-4718-4e89-82a9-b39e93a2220d","Type":"ContainerStarted","Data":"4a2d3396650367a2ad1736b003b1232bacccfcf75232c99b18bf87c4760a06bc"} Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.479784 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-9qfcx" event={"ID":"db58dfb9-4fd9-4896-add0-116a1ed9079d","Type":"ContainerStarted","Data":"f130aeea80b38efa526fe6fae1d7517e09010fc42ea9eaaa89e9d2e05904c641"} Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.485185 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" event={"ID":"2c9804b5-0729-4059-bb00-35fd1d355b3c","Type":"ContainerStarted","Data":"6ceefce5bc10f1c95777ae3f15fa0ced244c95d5367e45e7732d0eb1a11d9757"} Oct 08 09:22:03 crc kubenswrapper[4744]: I1008 09:22:03.595547 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l"] Oct 08 09:22:03 crc kubenswrapper[4744]: W1008 09:22:03.604508 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89b615a3_0a5d_483c_8d41_2dee267f1686.slice/crio-039a580f84fd88cd08011074b4d3d00814cd250bc0bd145fc575542ca91baefa WatchSource:0}: Error finding container 039a580f84fd88cd08011074b4d3d00814cd250bc0bd145fc575542ca91baefa: Status 404 returned error can't find the container with id 039a580f84fd88cd08011074b4d3d00814cd250bc0bd145fc575542ca91baefa Oct 08 09:22:04 crc kubenswrapper[4744]: I1008 09:22:04.496050 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" event={"ID":"89b615a3-0a5d-483c-8d41-2dee267f1686","Type":"ContainerStarted","Data":"039a580f84fd88cd08011074b4d3d00814cd250bc0bd145fc575542ca91baefa"} Oct 08 09:22:04 crc kubenswrapper[4744]: I1008 09:22:04.497701 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7cbf967b4c-hpjvs" event={"ID":"bb9f4b9c-410e-4b88-8172-f65d81b767f3","Type":"ContainerStarted","Data":"e398c4f05256069f834c11ac2b6c9e99a3f4ec62c0163409f66f5d679420372d"} Oct 08 09:22:04 crc kubenswrapper[4744]: I1008 09:22:04.497723 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7cbf967b4c-hpjvs" event={"ID":"bb9f4b9c-410e-4b88-8172-f65d81b767f3","Type":"ContainerStarted","Data":"3ec6e058ccaac3abc0a039c5020af2f7a8b14a4631d6420d9b7eed080729e409"} Oct 08 09:22:04 crc kubenswrapper[4744]: I1008 09:22:04.522890 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7cbf967b4c-hpjvs" podStartSLOduration=2.522859426 podStartE2EDuration="2.522859426s" podCreationTimestamp="2025-10-08 09:22:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:22:04.522783933 +0000 UTC m=+619.770429192" watchObservedRunningTime="2025-10-08 09:22:04.522859426 +0000 UTC m=+619.770504665" Oct 08 09:22:06 crc kubenswrapper[4744]: I1008 09:22:06.514047 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-tk2k5" event={"ID":"8b5fc549-4718-4e89-82a9-b39e93a2220d","Type":"ContainerStarted","Data":"d20048e547417e7b42249953a750db3bb6486f9a60fd607d5589a775f0f1b68d"} Oct 08 09:22:06 crc kubenswrapper[4744]: I1008 09:22:06.514332 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:06 crc kubenswrapper[4744]: I1008 09:22:06.516572 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-9qfcx" event={"ID":"db58dfb9-4fd9-4896-add0-116a1ed9079d","Type":"ContainerStarted","Data":"4fa5a2587f30126363ecb624125dc037c383f616b87df11d747188bdd3082591"} Oct 08 09:22:06 crc kubenswrapper[4744]: I1008 09:22:06.518320 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" event={"ID":"89b615a3-0a5d-483c-8d41-2dee267f1686","Type":"ContainerStarted","Data":"cdb475d6b8e7c0a37e84efaf3abcaa242e66fbdc68ee3f5ca6793e885012e7d1"} Oct 08 09:22:06 crc kubenswrapper[4744]: I1008 09:22:06.518504 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" Oct 08 09:22:06 crc kubenswrapper[4744]: I1008 09:22:06.522222 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" event={"ID":"2c9804b5-0729-4059-bb00-35fd1d355b3c","Type":"ContainerStarted","Data":"e315086dbd0ef3f76dcf37bbe0c78b80bf497b34347e3b6639cd535890e9b04c"} Oct 08 09:22:06 crc kubenswrapper[4744]: I1008 09:22:06.529497 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-tk2k5" podStartSLOduration=1.211961434 podStartE2EDuration="4.529474145s" podCreationTimestamp="2025-10-08 09:22:02 +0000 UTC" firstStartedPulling="2025-10-08 09:22:02.83507779 +0000 UTC m=+618.082723019" lastFinishedPulling="2025-10-08 09:22:06.152590491 +0000 UTC m=+621.400235730" observedRunningTime="2025-10-08 09:22:06.527841348 +0000 UTC m=+621.775486587" watchObservedRunningTime="2025-10-08 09:22:06.529474145 +0000 UTC m=+621.777119394" Oct 08 09:22:06 crc kubenswrapper[4744]: I1008 09:22:06.547205 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" podStartSLOduration=1.9612880129999999 podStartE2EDuration="4.547182638s" podCreationTimestamp="2025-10-08 09:22:02 +0000 UTC" firstStartedPulling="2025-10-08 09:22:03.606958852 +0000 UTC m=+618.854604091" lastFinishedPulling="2025-10-08 09:22:06.192853477 +0000 UTC m=+621.440498716" observedRunningTime="2025-10-08 09:22:06.543145601 +0000 UTC m=+621.790790860" watchObservedRunningTime="2025-10-08 09:22:06.547182638 +0000 UTC m=+621.794827877" Oct 08 09:22:06 crc kubenswrapper[4744]: I1008 09:22:06.575091 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-hrjff" podStartSLOduration=1.659558937 podStartE2EDuration="4.575073526s" podCreationTimestamp="2025-10-08 09:22:02 +0000 UTC" firstStartedPulling="2025-10-08 09:22:03.24115693 +0000 UTC m=+618.488802169" lastFinishedPulling="2025-10-08 09:22:06.156671519 +0000 UTC m=+621.404316758" observedRunningTime="2025-10-08 09:22:06.570112892 +0000 UTC m=+621.817758151" watchObservedRunningTime="2025-10-08 09:22:06.575073526 +0000 UTC m=+621.822718765" Oct 08 09:22:12 crc kubenswrapper[4744]: I1008 09:22:12.849835 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-tk2k5" Oct 08 09:22:13 crc kubenswrapper[4744]: I1008 09:22:13.253519 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:13 crc kubenswrapper[4744]: I1008 09:22:13.253607 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:13 crc kubenswrapper[4744]: I1008 09:22:13.261207 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:13 crc kubenswrapper[4744]: I1008 09:22:13.584764 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7cbf967b4c-hpjvs" Oct 08 09:22:13 crc kubenswrapper[4744]: I1008 09:22:13.694407 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5kssv"] Oct 08 09:22:23 crc kubenswrapper[4744]: I1008 09:22:23.365546 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-pnz4l" Oct 08 09:22:27 crc kubenswrapper[4744]: I1008 09:22:27.687555 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-9qfcx" event={"ID":"db58dfb9-4fd9-4896-add0-116a1ed9079d","Type":"ContainerStarted","Data":"5c9ad24125ea413c0847df55c852aba832b0b43e272c64ebb4d45b966ce92809"} Oct 08 09:22:27 crc kubenswrapper[4744]: I1008 09:22:27.714157 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-9qfcx" podStartSLOduration=1.731091387 podStartE2EDuration="25.714130427s" podCreationTimestamp="2025-10-08 09:22:02 +0000 UTC" firstStartedPulling="2025-10-08 09:22:03.125839789 +0000 UTC m=+618.373485028" lastFinishedPulling="2025-10-08 09:22:27.108878829 +0000 UTC m=+642.356524068" observedRunningTime="2025-10-08 09:22:27.708505814 +0000 UTC m=+642.956151073" watchObservedRunningTime="2025-10-08 09:22:27.714130427 +0000 UTC m=+642.961775676" Oct 08 09:22:36 crc kubenswrapper[4744]: I1008 09:22:36.830655 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp"] Oct 08 09:22:36 crc kubenswrapper[4744]: I1008 09:22:36.833029 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:36 crc kubenswrapper[4744]: I1008 09:22:36.835788 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 09:22:36 crc kubenswrapper[4744]: I1008 09:22:36.874810 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp"] Oct 08 09:22:36 crc kubenswrapper[4744]: I1008 09:22:36.955893 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efc6b240-277b-494e-9189-afd42c5a2aba-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp\" (UID: \"efc6b240-277b-494e-9189-afd42c5a2aba\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:36 crc kubenswrapper[4744]: I1008 09:22:36.955970 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efc6b240-277b-494e-9189-afd42c5a2aba-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp\" (UID: \"efc6b240-277b-494e-9189-afd42c5a2aba\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:36 crc kubenswrapper[4744]: I1008 09:22:36.956015 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcxkp\" (UniqueName: \"kubernetes.io/projected/efc6b240-277b-494e-9189-afd42c5a2aba-kube-api-access-zcxkp\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp\" (UID: \"efc6b240-277b-494e-9189-afd42c5a2aba\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:37 crc kubenswrapper[4744]: I1008 09:22:37.057746 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcxkp\" (UniqueName: \"kubernetes.io/projected/efc6b240-277b-494e-9189-afd42c5a2aba-kube-api-access-zcxkp\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp\" (UID: \"efc6b240-277b-494e-9189-afd42c5a2aba\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:37 crc kubenswrapper[4744]: I1008 09:22:37.057881 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efc6b240-277b-494e-9189-afd42c5a2aba-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp\" (UID: \"efc6b240-277b-494e-9189-afd42c5a2aba\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:37 crc kubenswrapper[4744]: I1008 09:22:37.057913 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efc6b240-277b-494e-9189-afd42c5a2aba-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp\" (UID: \"efc6b240-277b-494e-9189-afd42c5a2aba\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:37 crc kubenswrapper[4744]: I1008 09:22:37.058512 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efc6b240-277b-494e-9189-afd42c5a2aba-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp\" (UID: \"efc6b240-277b-494e-9189-afd42c5a2aba\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:37 crc kubenswrapper[4744]: I1008 09:22:37.058715 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efc6b240-277b-494e-9189-afd42c5a2aba-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp\" (UID: \"efc6b240-277b-494e-9189-afd42c5a2aba\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:37 crc kubenswrapper[4744]: I1008 09:22:37.076702 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcxkp\" (UniqueName: \"kubernetes.io/projected/efc6b240-277b-494e-9189-afd42c5a2aba-kube-api-access-zcxkp\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp\" (UID: \"efc6b240-277b-494e-9189-afd42c5a2aba\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:37 crc kubenswrapper[4744]: I1008 09:22:37.197787 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:37 crc kubenswrapper[4744]: I1008 09:22:37.412725 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp"] Oct 08 09:22:37 crc kubenswrapper[4744]: I1008 09:22:37.748931 4744 generic.go:334] "Generic (PLEG): container finished" podID="efc6b240-277b-494e-9189-afd42c5a2aba" containerID="a1e8ac8976ceb0aeef5e8b65e24c0c3c50e3b874110391c0c19e31f43dbc660f" exitCode=0 Oct 08 09:22:37 crc kubenswrapper[4744]: I1008 09:22:37.749181 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" event={"ID":"efc6b240-277b-494e-9189-afd42c5a2aba","Type":"ContainerDied","Data":"a1e8ac8976ceb0aeef5e8b65e24c0c3c50e3b874110391c0c19e31f43dbc660f"} Oct 08 09:22:37 crc kubenswrapper[4744]: I1008 09:22:37.749206 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" event={"ID":"efc6b240-277b-494e-9189-afd42c5a2aba","Type":"ContainerStarted","Data":"46a0a3dc13dc1289bb9e5b219ff947d6f57b723028859afc7823153bccfcffc1"} Oct 08 09:22:38 crc kubenswrapper[4744]: I1008 09:22:38.769158 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-5kssv" podUID="bb805699-c562-4a60-a45a-67dbbba49779" containerName="console" containerID="cri-o://c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588" gracePeriod=15 Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.343973 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5kssv_bb805699-c562-4a60-a45a-67dbbba49779/console/0.log" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.344699 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.388931 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-service-ca\") pod \"bb805699-c562-4a60-a45a-67dbbba49779\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.389143 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb805699-c562-4a60-a45a-67dbbba49779-console-serving-cert\") pod \"bb805699-c562-4a60-a45a-67dbbba49779\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.389247 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk8mt\" (UniqueName: \"kubernetes.io/projected/bb805699-c562-4a60-a45a-67dbbba49779-kube-api-access-vk8mt\") pod \"bb805699-c562-4a60-a45a-67dbbba49779\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.389349 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-trusted-ca-bundle\") pod \"bb805699-c562-4a60-a45a-67dbbba49779\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.389465 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-console-config\") pod \"bb805699-c562-4a60-a45a-67dbbba49779\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.389577 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-oauth-serving-cert\") pod \"bb805699-c562-4a60-a45a-67dbbba49779\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.389652 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb805699-c562-4a60-a45a-67dbbba49779-console-oauth-config\") pod \"bb805699-c562-4a60-a45a-67dbbba49779\" (UID: \"bb805699-c562-4a60-a45a-67dbbba49779\") " Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.390545 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-console-config" (OuterVolumeSpecName: "console-config") pod "bb805699-c562-4a60-a45a-67dbbba49779" (UID: "bb805699-c562-4a60-a45a-67dbbba49779"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.390760 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "bb805699-c562-4a60-a45a-67dbbba49779" (UID: "bb805699-c562-4a60-a45a-67dbbba49779"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.391045 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "bb805699-c562-4a60-a45a-67dbbba49779" (UID: "bb805699-c562-4a60-a45a-67dbbba49779"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.391121 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-service-ca" (OuterVolumeSpecName: "service-ca") pod "bb805699-c562-4a60-a45a-67dbbba49779" (UID: "bb805699-c562-4a60-a45a-67dbbba49779"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.396787 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb805699-c562-4a60-a45a-67dbbba49779-kube-api-access-vk8mt" (OuterVolumeSpecName: "kube-api-access-vk8mt") pod "bb805699-c562-4a60-a45a-67dbbba49779" (UID: "bb805699-c562-4a60-a45a-67dbbba49779"). InnerVolumeSpecName "kube-api-access-vk8mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.397005 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb805699-c562-4a60-a45a-67dbbba49779-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "bb805699-c562-4a60-a45a-67dbbba49779" (UID: "bb805699-c562-4a60-a45a-67dbbba49779"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.397184 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb805699-c562-4a60-a45a-67dbbba49779-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "bb805699-c562-4a60-a45a-67dbbba49779" (UID: "bb805699-c562-4a60-a45a-67dbbba49779"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.491704 4744 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.491858 4744 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb805699-c562-4a60-a45a-67dbbba49779-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.491932 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.492027 4744 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb805699-c562-4a60-a45a-67dbbba49779-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.492146 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk8mt\" (UniqueName: \"kubernetes.io/projected/bb805699-c562-4a60-a45a-67dbbba49779-kube-api-access-vk8mt\") on node \"crc\" DevicePath \"\"" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.492249 4744 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.492472 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb805699-c562-4a60-a45a-67dbbba49779-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.769162 4744 generic.go:334] "Generic (PLEG): container finished" podID="efc6b240-277b-494e-9189-afd42c5a2aba" containerID="9f9701a83f662661c94048f5d404b422cdf67f3fb50b7e2f234f16452ebdf6f6" exitCode=0 Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.769727 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" event={"ID":"efc6b240-277b-494e-9189-afd42c5a2aba","Type":"ContainerDied","Data":"9f9701a83f662661c94048f5d404b422cdf67f3fb50b7e2f234f16452ebdf6f6"} Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.778240 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5kssv_bb805699-c562-4a60-a45a-67dbbba49779/console/0.log" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.778307 4744 generic.go:334] "Generic (PLEG): container finished" podID="bb805699-c562-4a60-a45a-67dbbba49779" containerID="c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588" exitCode=2 Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.778346 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5kssv" event={"ID":"bb805699-c562-4a60-a45a-67dbbba49779","Type":"ContainerDied","Data":"c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588"} Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.778397 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5kssv" event={"ID":"bb805699-c562-4a60-a45a-67dbbba49779","Type":"ContainerDied","Data":"4656b1c3d2f280cb69f6fd396de0bdddb2cf40d182e65b95c2d905ea803cb303"} Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.778422 4744 scope.go:117] "RemoveContainer" containerID="c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.778562 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5kssv" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.811920 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5kssv"] Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.811985 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-5kssv"] Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.823644 4744 scope.go:117] "RemoveContainer" containerID="c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588" Oct 08 09:22:39 crc kubenswrapper[4744]: E1008 09:22:39.824423 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588\": container with ID starting with c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588 not found: ID does not exist" containerID="c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588" Oct 08 09:22:39 crc kubenswrapper[4744]: I1008 09:22:39.824492 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588"} err="failed to get container status \"c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588\": rpc error: code = NotFound desc = could not find container \"c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588\": container with ID starting with c35867661b29d560f23de4bee942e074dd3f653e53e70faec2fec37b36fe2588 not found: ID does not exist" Oct 08 09:22:40 crc kubenswrapper[4744]: I1008 09:22:40.791549 4744 generic.go:334] "Generic (PLEG): container finished" podID="efc6b240-277b-494e-9189-afd42c5a2aba" containerID="d761af03b2656bbb51d4817aec4fd71b9357bbad302e836cf0502ce1d7436a99" exitCode=0 Oct 08 09:22:40 crc kubenswrapper[4744]: I1008 09:22:40.791601 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" event={"ID":"efc6b240-277b-494e-9189-afd42c5a2aba","Type":"ContainerDied","Data":"d761af03b2656bbb51d4817aec4fd71b9357bbad302e836cf0502ce1d7436a99"} Oct 08 09:22:41 crc kubenswrapper[4744]: I1008 09:22:41.463990 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb805699-c562-4a60-a45a-67dbbba49779" path="/var/lib/kubelet/pods/bb805699-c562-4a60-a45a-67dbbba49779/volumes" Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.044154 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.129944 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcxkp\" (UniqueName: \"kubernetes.io/projected/efc6b240-277b-494e-9189-afd42c5a2aba-kube-api-access-zcxkp\") pod \"efc6b240-277b-494e-9189-afd42c5a2aba\" (UID: \"efc6b240-277b-494e-9189-afd42c5a2aba\") " Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.130105 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efc6b240-277b-494e-9189-afd42c5a2aba-util\") pod \"efc6b240-277b-494e-9189-afd42c5a2aba\" (UID: \"efc6b240-277b-494e-9189-afd42c5a2aba\") " Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.130212 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efc6b240-277b-494e-9189-afd42c5a2aba-bundle\") pod \"efc6b240-277b-494e-9189-afd42c5a2aba\" (UID: \"efc6b240-277b-494e-9189-afd42c5a2aba\") " Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.131885 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efc6b240-277b-494e-9189-afd42c5a2aba-bundle" (OuterVolumeSpecName: "bundle") pod "efc6b240-277b-494e-9189-afd42c5a2aba" (UID: "efc6b240-277b-494e-9189-afd42c5a2aba"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.146718 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efc6b240-277b-494e-9189-afd42c5a2aba-kube-api-access-zcxkp" (OuterVolumeSpecName: "kube-api-access-zcxkp") pod "efc6b240-277b-494e-9189-afd42c5a2aba" (UID: "efc6b240-277b-494e-9189-afd42c5a2aba"). InnerVolumeSpecName "kube-api-access-zcxkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.157308 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efc6b240-277b-494e-9189-afd42c5a2aba-util" (OuterVolumeSpecName: "util") pod "efc6b240-277b-494e-9189-afd42c5a2aba" (UID: "efc6b240-277b-494e-9189-afd42c5a2aba"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.231215 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcxkp\" (UniqueName: \"kubernetes.io/projected/efc6b240-277b-494e-9189-afd42c5a2aba-kube-api-access-zcxkp\") on node \"crc\" DevicePath \"\"" Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.231262 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efc6b240-277b-494e-9189-afd42c5a2aba-util\") on node \"crc\" DevicePath \"\"" Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.231274 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efc6b240-277b-494e-9189-afd42c5a2aba-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.810204 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" event={"ID":"efc6b240-277b-494e-9189-afd42c5a2aba","Type":"ContainerDied","Data":"46a0a3dc13dc1289bb9e5b219ff947d6f57b723028859afc7823153bccfcffc1"} Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.810627 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46a0a3dc13dc1289bb9e5b219ff947d6f57b723028859afc7823153bccfcffc1" Oct 08 09:22:42 crc kubenswrapper[4744]: I1008 09:22:42.810560 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.408320 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj"] Oct 08 09:22:51 crc kubenswrapper[4744]: E1008 09:22:51.409121 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb805699-c562-4a60-a45a-67dbbba49779" containerName="console" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.409136 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb805699-c562-4a60-a45a-67dbbba49779" containerName="console" Oct 08 09:22:51 crc kubenswrapper[4744]: E1008 09:22:51.409154 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efc6b240-277b-494e-9189-afd42c5a2aba" containerName="extract" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.409160 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="efc6b240-277b-494e-9189-afd42c5a2aba" containerName="extract" Oct 08 09:22:51 crc kubenswrapper[4744]: E1008 09:22:51.409169 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efc6b240-277b-494e-9189-afd42c5a2aba" containerName="pull" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.409175 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="efc6b240-277b-494e-9189-afd42c5a2aba" containerName="pull" Oct 08 09:22:51 crc kubenswrapper[4744]: E1008 09:22:51.409182 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efc6b240-277b-494e-9189-afd42c5a2aba" containerName="util" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.409188 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="efc6b240-277b-494e-9189-afd42c5a2aba" containerName="util" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.409280 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb805699-c562-4a60-a45a-67dbbba49779" containerName="console" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.409289 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="efc6b240-277b-494e-9189-afd42c5a2aba" containerName="extract" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.409689 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.415017 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.417614 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.417824 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.417949 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-lt4bm" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.424317 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.440867 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj"] Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.483637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpzh4\" (UniqueName: \"kubernetes.io/projected/fb79b0ea-df0c-4f82-a297-4ce347bdc363-kube-api-access-tpzh4\") pod \"metallb-operator-controller-manager-5998774b9-gk4vj\" (UID: \"fb79b0ea-df0c-4f82-a297-4ce347bdc363\") " pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.483710 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fb79b0ea-df0c-4f82-a297-4ce347bdc363-webhook-cert\") pod \"metallb-operator-controller-manager-5998774b9-gk4vj\" (UID: \"fb79b0ea-df0c-4f82-a297-4ce347bdc363\") " pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.483803 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fb79b0ea-df0c-4f82-a297-4ce347bdc363-apiservice-cert\") pod \"metallb-operator-controller-manager-5998774b9-gk4vj\" (UID: \"fb79b0ea-df0c-4f82-a297-4ce347bdc363\") " pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.584231 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fb79b0ea-df0c-4f82-a297-4ce347bdc363-webhook-cert\") pod \"metallb-operator-controller-manager-5998774b9-gk4vj\" (UID: \"fb79b0ea-df0c-4f82-a297-4ce347bdc363\") " pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.584307 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fb79b0ea-df0c-4f82-a297-4ce347bdc363-apiservice-cert\") pod \"metallb-operator-controller-manager-5998774b9-gk4vj\" (UID: \"fb79b0ea-df0c-4f82-a297-4ce347bdc363\") " pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.584337 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpzh4\" (UniqueName: \"kubernetes.io/projected/fb79b0ea-df0c-4f82-a297-4ce347bdc363-kube-api-access-tpzh4\") pod \"metallb-operator-controller-manager-5998774b9-gk4vj\" (UID: \"fb79b0ea-df0c-4f82-a297-4ce347bdc363\") " pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.595063 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fb79b0ea-df0c-4f82-a297-4ce347bdc363-webhook-cert\") pod \"metallb-operator-controller-manager-5998774b9-gk4vj\" (UID: \"fb79b0ea-df0c-4f82-a297-4ce347bdc363\") " pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.602991 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fb79b0ea-df0c-4f82-a297-4ce347bdc363-apiservice-cert\") pod \"metallb-operator-controller-manager-5998774b9-gk4vj\" (UID: \"fb79b0ea-df0c-4f82-a297-4ce347bdc363\") " pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.628212 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpzh4\" (UniqueName: \"kubernetes.io/projected/fb79b0ea-df0c-4f82-a297-4ce347bdc363-kube-api-access-tpzh4\") pod \"metallb-operator-controller-manager-5998774b9-gk4vj\" (UID: \"fb79b0ea-df0c-4f82-a297-4ce347bdc363\") " pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.724541 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.794482 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s"] Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.795197 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.799621 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.800035 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-g4t8d" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.800320 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.817552 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s"] Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.988858 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msssl\" (UniqueName: \"kubernetes.io/projected/9de92423-8f60-43bd-842c-b47f51c9e20c-kube-api-access-msssl\") pod \"metallb-operator-webhook-server-dbd787fb9-lsv2s\" (UID: \"9de92423-8f60-43bd-842c-b47f51c9e20c\") " pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.988920 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9de92423-8f60-43bd-842c-b47f51c9e20c-webhook-cert\") pod \"metallb-operator-webhook-server-dbd787fb9-lsv2s\" (UID: \"9de92423-8f60-43bd-842c-b47f51c9e20c\") " pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:22:51 crc kubenswrapper[4744]: I1008 09:22:51.988976 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9de92423-8f60-43bd-842c-b47f51c9e20c-apiservice-cert\") pod \"metallb-operator-webhook-server-dbd787fb9-lsv2s\" (UID: \"9de92423-8f60-43bd-842c-b47f51c9e20c\") " pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:22:52 crc kubenswrapper[4744]: I1008 09:22:52.089917 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msssl\" (UniqueName: \"kubernetes.io/projected/9de92423-8f60-43bd-842c-b47f51c9e20c-kube-api-access-msssl\") pod \"metallb-operator-webhook-server-dbd787fb9-lsv2s\" (UID: \"9de92423-8f60-43bd-842c-b47f51c9e20c\") " pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:22:52 crc kubenswrapper[4744]: I1008 09:22:52.089961 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9de92423-8f60-43bd-842c-b47f51c9e20c-webhook-cert\") pod \"metallb-operator-webhook-server-dbd787fb9-lsv2s\" (UID: \"9de92423-8f60-43bd-842c-b47f51c9e20c\") " pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:22:52 crc kubenswrapper[4744]: I1008 09:22:52.089983 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9de92423-8f60-43bd-842c-b47f51c9e20c-apiservice-cert\") pod \"metallb-operator-webhook-server-dbd787fb9-lsv2s\" (UID: \"9de92423-8f60-43bd-842c-b47f51c9e20c\") " pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:22:52 crc kubenswrapper[4744]: I1008 09:22:52.096135 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9de92423-8f60-43bd-842c-b47f51c9e20c-webhook-cert\") pod \"metallb-operator-webhook-server-dbd787fb9-lsv2s\" (UID: \"9de92423-8f60-43bd-842c-b47f51c9e20c\") " pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:22:52 crc kubenswrapper[4744]: I1008 09:22:52.097080 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9de92423-8f60-43bd-842c-b47f51c9e20c-apiservice-cert\") pod \"metallb-operator-webhook-server-dbd787fb9-lsv2s\" (UID: \"9de92423-8f60-43bd-842c-b47f51c9e20c\") " pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:22:52 crc kubenswrapper[4744]: I1008 09:22:52.114279 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msssl\" (UniqueName: \"kubernetes.io/projected/9de92423-8f60-43bd-842c-b47f51c9e20c-kube-api-access-msssl\") pod \"metallb-operator-webhook-server-dbd787fb9-lsv2s\" (UID: \"9de92423-8f60-43bd-842c-b47f51c9e20c\") " pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:22:52 crc kubenswrapper[4744]: I1008 09:22:52.214026 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj"] Oct 08 09:22:52 crc kubenswrapper[4744]: W1008 09:22:52.225119 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb79b0ea_df0c_4f82_a297_4ce347bdc363.slice/crio-587a3684af2585d88bdeec0030d2f61a184d01ba40cd5c5c067ea668f61705e7 WatchSource:0}: Error finding container 587a3684af2585d88bdeec0030d2f61a184d01ba40cd5c5c067ea668f61705e7: Status 404 returned error can't find the container with id 587a3684af2585d88bdeec0030d2f61a184d01ba40cd5c5c067ea668f61705e7 Oct 08 09:22:52 crc kubenswrapper[4744]: I1008 09:22:52.410679 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:22:52 crc kubenswrapper[4744]: I1008 09:22:52.692062 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s"] Oct 08 09:22:52 crc kubenswrapper[4744]: W1008 09:22:52.717535 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9de92423_8f60_43bd_842c_b47f51c9e20c.slice/crio-99f0ab17a9623fd34200ec6a3a4aa93869ed7cefb26b1a104a0bf1cbf19a16cd WatchSource:0}: Error finding container 99f0ab17a9623fd34200ec6a3a4aa93869ed7cefb26b1a104a0bf1cbf19a16cd: Status 404 returned error can't find the container with id 99f0ab17a9623fd34200ec6a3a4aa93869ed7cefb26b1a104a0bf1cbf19a16cd Oct 08 09:22:52 crc kubenswrapper[4744]: I1008 09:22:52.879819 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" event={"ID":"fb79b0ea-df0c-4f82-a297-4ce347bdc363","Type":"ContainerStarted","Data":"587a3684af2585d88bdeec0030d2f61a184d01ba40cd5c5c067ea668f61705e7"} Oct 08 09:22:52 crc kubenswrapper[4744]: I1008 09:22:52.881872 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" event={"ID":"9de92423-8f60-43bd-842c-b47f51c9e20c","Type":"ContainerStarted","Data":"99f0ab17a9623fd34200ec6a3a4aa93869ed7cefb26b1a104a0bf1cbf19a16cd"} Oct 08 09:22:56 crc kubenswrapper[4744]: I1008 09:22:56.910153 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" event={"ID":"fb79b0ea-df0c-4f82-a297-4ce347bdc363","Type":"ContainerStarted","Data":"7cd3e981bb53e92f090ad20ea38074c88f01cd26ac48339fe1d4aea1d63dafdf"} Oct 08 09:22:56 crc kubenswrapper[4744]: I1008 09:22:56.910723 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:22:56 crc kubenswrapper[4744]: I1008 09:22:56.947429 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" podStartSLOduration=2.131280729 podStartE2EDuration="5.947380649s" podCreationTimestamp="2025-10-08 09:22:51 +0000 UTC" firstStartedPulling="2025-10-08 09:22:52.227967824 +0000 UTC m=+667.475613063" lastFinishedPulling="2025-10-08 09:22:56.044067754 +0000 UTC m=+671.291712983" observedRunningTime="2025-10-08 09:22:56.93905765 +0000 UTC m=+672.186702929" watchObservedRunningTime="2025-10-08 09:22:56.947380649 +0000 UTC m=+672.195025898" Oct 08 09:23:01 crc kubenswrapper[4744]: I1008 09:23:01.958814 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" event={"ID":"9de92423-8f60-43bd-842c-b47f51c9e20c","Type":"ContainerStarted","Data":"0d5d6961b1be837efa9f3c1a392d43962b564d27917a04e33aead570f412e9de"} Oct 08 09:23:01 crc kubenswrapper[4744]: I1008 09:23:01.960304 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:23:01 crc kubenswrapper[4744]: I1008 09:23:01.980807 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" podStartSLOduration=2.083498614 podStartE2EDuration="10.980780926s" podCreationTimestamp="2025-10-08 09:22:51 +0000 UTC" firstStartedPulling="2025-10-08 09:22:52.720146282 +0000 UTC m=+667.967791521" lastFinishedPulling="2025-10-08 09:23:01.617428594 +0000 UTC m=+676.865073833" observedRunningTime="2025-10-08 09:23:01.978976644 +0000 UTC m=+677.226621903" watchObservedRunningTime="2025-10-08 09:23:01.980780926 +0000 UTC m=+677.228426165" Oct 08 09:23:12 crc kubenswrapper[4744]: I1008 09:23:12.420894 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-dbd787fb9-lsv2s" Oct 08 09:23:19 crc kubenswrapper[4744]: I1008 09:23:19.690993 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:23:19 crc kubenswrapper[4744]: I1008 09:23:19.691519 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:23:31 crc kubenswrapper[4744]: I1008 09:23:31.729420 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5998774b9-gk4vj" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.559406 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-5skh4"] Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.561815 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.564523 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h"] Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.565248 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.567198 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.567246 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.569820 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-zpk24" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.570009 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.617104 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h"] Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.633026 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6jxk\" (UniqueName: \"kubernetes.io/projected/cf7d04b2-1136-4427-8939-e2974337725c-kube-api-access-m6jxk\") pod \"frr-k8s-webhook-server-64bf5d555-ptt9h\" (UID: \"cf7d04b2-1136-4427-8939-e2974337725c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.633076 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/38b9b97b-50ef-48c1-b917-a6c2f2be304b-frr-sockets\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.633306 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/38b9b97b-50ef-48c1-b917-a6c2f2be304b-frr-conf\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.633354 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/38b9b97b-50ef-48c1-b917-a6c2f2be304b-frr-startup\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.633401 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf7d04b2-1136-4427-8939-e2974337725c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-ptt9h\" (UID: \"cf7d04b2-1136-4427-8939-e2974337725c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.633446 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38b9b97b-50ef-48c1-b917-a6c2f2be304b-metrics-certs\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.633479 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/38b9b97b-50ef-48c1-b917-a6c2f2be304b-reloader\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.633494 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/38b9b97b-50ef-48c1-b917-a6c2f2be304b-metrics\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.633511 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcsnx\" (UniqueName: \"kubernetes.io/projected/38b9b97b-50ef-48c1-b917-a6c2f2be304b-kube-api-access-wcsnx\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.702208 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-89qxn"] Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.703493 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-89qxn" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.705503 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-8rkml"] Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.706821 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.710814 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.714593 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-t7rz9" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.714799 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.725609 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-8rkml"] Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.727645 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.727673 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.734455 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-memberlist\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.735067 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/38b9b97b-50ef-48c1-b917-a6c2f2be304b-reloader\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.735153 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/38b9b97b-50ef-48c1-b917-a6c2f2be304b-metrics\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.735222 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcsnx\" (UniqueName: \"kubernetes.io/projected/38b9b97b-50ef-48c1-b917-a6c2f2be304b-kube-api-access-wcsnx\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.735311 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6jxk\" (UniqueName: \"kubernetes.io/projected/cf7d04b2-1136-4427-8939-e2974337725c-kube-api-access-m6jxk\") pod \"frr-k8s-webhook-server-64bf5d555-ptt9h\" (UID: \"cf7d04b2-1136-4427-8939-e2974337725c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.735397 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f0fc93a-5521-4a7a-90d7-7b1794dc728f-cert\") pod \"controller-68d546b9d8-8rkml\" (UID: \"4f0fc93a-5521-4a7a-90d7-7b1794dc728f\") " pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.735478 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/38b9b97b-50ef-48c1-b917-a6c2f2be304b-frr-sockets\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.735554 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-metrics-certs\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.735632 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/38b9b97b-50ef-48c1-b917-a6c2f2be304b-frr-conf\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.735725 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp4cq\" (UniqueName: \"kubernetes.io/projected/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-kube-api-access-tp4cq\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.735819 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/38b9b97b-50ef-48c1-b917-a6c2f2be304b-frr-startup\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.735911 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f0fc93a-5521-4a7a-90d7-7b1794dc728f-metrics-certs\") pod \"controller-68d546b9d8-8rkml\" (UID: \"4f0fc93a-5521-4a7a-90d7-7b1794dc728f\") " pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.736016 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf7d04b2-1136-4427-8939-e2974337725c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-ptt9h\" (UID: \"cf7d04b2-1136-4427-8939-e2974337725c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.736130 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-metallb-excludel2\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.736249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38b9b97b-50ef-48c1-b917-a6c2f2be304b-metrics-certs\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.736321 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgxqx\" (UniqueName: \"kubernetes.io/projected/4f0fc93a-5521-4a7a-90d7-7b1794dc728f-kube-api-access-jgxqx\") pod \"controller-68d546b9d8-8rkml\" (UID: \"4f0fc93a-5521-4a7a-90d7-7b1794dc728f\") " pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.736794 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/38b9b97b-50ef-48c1-b917-a6c2f2be304b-frr-conf\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.736878 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/38b9b97b-50ef-48c1-b917-a6c2f2be304b-reloader\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: E1008 09:23:32.736991 4744 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 08 09:23:32 crc kubenswrapper[4744]: E1008 09:23:32.737041 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf7d04b2-1136-4427-8939-e2974337725c-cert podName:cf7d04b2-1136-4427-8939-e2974337725c nodeName:}" failed. No retries permitted until 2025-10-08 09:23:33.237025524 +0000 UTC m=+708.484670763 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cf7d04b2-1136-4427-8939-e2974337725c-cert") pod "frr-k8s-webhook-server-64bf5d555-ptt9h" (UID: "cf7d04b2-1136-4427-8939-e2974337725c") : secret "frr-k8s-webhook-server-cert" not found Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.737333 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/38b9b97b-50ef-48c1-b917-a6c2f2be304b-metrics\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: E1008 09:23:32.737482 4744 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.737518 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/38b9b97b-50ef-48c1-b917-a6c2f2be304b-frr-sockets\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: E1008 09:23:32.737553 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/38b9b97b-50ef-48c1-b917-a6c2f2be304b-metrics-certs podName:38b9b97b-50ef-48c1-b917-a6c2f2be304b nodeName:}" failed. No retries permitted until 2025-10-08 09:23:33.237538979 +0000 UTC m=+708.485184208 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/38b9b97b-50ef-48c1-b917-a6c2f2be304b-metrics-certs") pod "frr-k8s-5skh4" (UID: "38b9b97b-50ef-48c1-b917-a6c2f2be304b") : secret "frr-k8s-certs-secret" not found Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.737685 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/38b9b97b-50ef-48c1-b917-a6c2f2be304b-frr-startup\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.778786 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6jxk\" (UniqueName: \"kubernetes.io/projected/cf7d04b2-1136-4427-8939-e2974337725c-kube-api-access-m6jxk\") pod \"frr-k8s-webhook-server-64bf5d555-ptt9h\" (UID: \"cf7d04b2-1136-4427-8939-e2974337725c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.783832 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcsnx\" (UniqueName: \"kubernetes.io/projected/38b9b97b-50ef-48c1-b917-a6c2f2be304b-kube-api-access-wcsnx\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.837352 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-memberlist\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.837438 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f0fc93a-5521-4a7a-90d7-7b1794dc728f-cert\") pod \"controller-68d546b9d8-8rkml\" (UID: \"4f0fc93a-5521-4a7a-90d7-7b1794dc728f\") " pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.837474 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-metrics-certs\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.837515 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp4cq\" (UniqueName: \"kubernetes.io/projected/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-kube-api-access-tp4cq\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.837543 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f0fc93a-5521-4a7a-90d7-7b1794dc728f-metrics-certs\") pod \"controller-68d546b9d8-8rkml\" (UID: \"4f0fc93a-5521-4a7a-90d7-7b1794dc728f\") " pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.837582 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-metallb-excludel2\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.837637 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgxqx\" (UniqueName: \"kubernetes.io/projected/4f0fc93a-5521-4a7a-90d7-7b1794dc728f-kube-api-access-jgxqx\") pod \"controller-68d546b9d8-8rkml\" (UID: \"4f0fc93a-5521-4a7a-90d7-7b1794dc728f\") " pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:32 crc kubenswrapper[4744]: E1008 09:23:32.838035 4744 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 09:23:32 crc kubenswrapper[4744]: E1008 09:23:32.838086 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-memberlist podName:349b3e3f-87d0-4e3d-b2dc-c96452a7199c nodeName:}" failed. No retries permitted until 2025-10-08 09:23:33.338069081 +0000 UTC m=+708.585714320 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-memberlist") pod "speaker-89qxn" (UID: "349b3e3f-87d0-4e3d-b2dc-c96452a7199c") : secret "metallb-memberlist" not found Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.840508 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-metallb-excludel2\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:32 crc kubenswrapper[4744]: E1008 09:23:32.840982 4744 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 08 09:23:32 crc kubenswrapper[4744]: E1008 09:23:32.841048 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-metrics-certs podName:349b3e3f-87d0-4e3d-b2dc-c96452a7199c nodeName:}" failed. No retries permitted until 2025-10-08 09:23:33.341025996 +0000 UTC m=+708.588671235 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-metrics-certs") pod "speaker-89qxn" (UID: "349b3e3f-87d0-4e3d-b2dc-c96452a7199c") : secret "speaker-certs-secret" not found Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.841361 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4f0fc93a-5521-4a7a-90d7-7b1794dc728f-cert\") pod \"controller-68d546b9d8-8rkml\" (UID: \"4f0fc93a-5521-4a7a-90d7-7b1794dc728f\") " pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.847359 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4f0fc93a-5521-4a7a-90d7-7b1794dc728f-metrics-certs\") pod \"controller-68d546b9d8-8rkml\" (UID: \"4f0fc93a-5521-4a7a-90d7-7b1794dc728f\") " pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.864921 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgxqx\" (UniqueName: \"kubernetes.io/projected/4f0fc93a-5521-4a7a-90d7-7b1794dc728f-kube-api-access-jgxqx\") pod \"controller-68d546b9d8-8rkml\" (UID: \"4f0fc93a-5521-4a7a-90d7-7b1794dc728f\") " pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:32 crc kubenswrapper[4744]: I1008 09:23:32.868757 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp4cq\" (UniqueName: \"kubernetes.io/projected/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-kube-api-access-tp4cq\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:33 crc kubenswrapper[4744]: I1008 09:23:33.028513 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:33 crc kubenswrapper[4744]: I1008 09:23:33.241635 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf7d04b2-1136-4427-8939-e2974337725c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-ptt9h\" (UID: \"cf7d04b2-1136-4427-8939-e2974337725c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" Oct 08 09:23:33 crc kubenswrapper[4744]: I1008 09:23:33.242037 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38b9b97b-50ef-48c1-b917-a6c2f2be304b-metrics-certs\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:33 crc kubenswrapper[4744]: I1008 09:23:33.246137 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/38b9b97b-50ef-48c1-b917-a6c2f2be304b-metrics-certs\") pod \"frr-k8s-5skh4\" (UID: \"38b9b97b-50ef-48c1-b917-a6c2f2be304b\") " pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:33 crc kubenswrapper[4744]: I1008 09:23:33.250527 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf7d04b2-1136-4427-8939-e2974337725c-cert\") pod \"frr-k8s-webhook-server-64bf5d555-ptt9h\" (UID: \"cf7d04b2-1136-4427-8939-e2974337725c\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" Oct 08 09:23:33 crc kubenswrapper[4744]: I1008 09:23:33.287090 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-8rkml"] Oct 08 09:23:33 crc kubenswrapper[4744]: I1008 09:23:33.344629 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-memberlist\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:33 crc kubenswrapper[4744]: I1008 09:23:33.344696 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-metrics-certs\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:33 crc kubenswrapper[4744]: E1008 09:23:33.345121 4744 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 09:23:33 crc kubenswrapper[4744]: E1008 09:23:33.345177 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-memberlist podName:349b3e3f-87d0-4e3d-b2dc-c96452a7199c nodeName:}" failed. No retries permitted until 2025-10-08 09:23:34.345162458 +0000 UTC m=+709.592807697 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-memberlist") pod "speaker-89qxn" (UID: "349b3e3f-87d0-4e3d-b2dc-c96452a7199c") : secret "metallb-memberlist" not found Oct 08 09:23:33 crc kubenswrapper[4744]: I1008 09:23:33.349010 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-metrics-certs\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:33 crc kubenswrapper[4744]: I1008 09:23:33.482214 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:33 crc kubenswrapper[4744]: I1008 09:23:33.533560 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" Oct 08 09:23:33 crc kubenswrapper[4744]: I1008 09:23:33.754671 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h"] Oct 08 09:23:33 crc kubenswrapper[4744]: W1008 09:23:33.759276 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf7d04b2_1136_4427_8939_e2974337725c.slice/crio-9d70703233f58a76b122d4e7da3047d6c55975eb9c915478ae00baa0d8499e00 WatchSource:0}: Error finding container 9d70703233f58a76b122d4e7da3047d6c55975eb9c915478ae00baa0d8499e00: Status 404 returned error can't find the container with id 9d70703233f58a76b122d4e7da3047d6c55975eb9c915478ae00baa0d8499e00 Oct 08 09:23:34 crc kubenswrapper[4744]: I1008 09:23:34.175768 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-8rkml" event={"ID":"4f0fc93a-5521-4a7a-90d7-7b1794dc728f","Type":"ContainerStarted","Data":"a09d94875e4ebd71c8ca41e024e46e18cebaf3509a1d675182b12bdde0a11186"} Oct 08 09:23:34 crc kubenswrapper[4744]: I1008 09:23:34.176211 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:34 crc kubenswrapper[4744]: I1008 09:23:34.176226 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-8rkml" event={"ID":"4f0fc93a-5521-4a7a-90d7-7b1794dc728f","Type":"ContainerStarted","Data":"5c66e32fb973fa03b5b943c52f5b1ec6f277e840737860047163dcd9f939d023"} Oct 08 09:23:34 crc kubenswrapper[4744]: I1008 09:23:34.176243 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-8rkml" event={"ID":"4f0fc93a-5521-4a7a-90d7-7b1794dc728f","Type":"ContainerStarted","Data":"96ab9c9dce8d8727a9c0d6297e3926349afb59df5a90f05d50a5b0e9783c6bf3"} Oct 08 09:23:34 crc kubenswrapper[4744]: I1008 09:23:34.177723 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" event={"ID":"cf7d04b2-1136-4427-8939-e2974337725c","Type":"ContainerStarted","Data":"9d70703233f58a76b122d4e7da3047d6c55975eb9c915478ae00baa0d8499e00"} Oct 08 09:23:34 crc kubenswrapper[4744]: I1008 09:23:34.178934 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5skh4" event={"ID":"38b9b97b-50ef-48c1-b917-a6c2f2be304b","Type":"ContainerStarted","Data":"c4bc3ad5e66daa685c13678351cf7ea6081aec32e495e149b2d759f93e6e6e6d"} Oct 08 09:23:34 crc kubenswrapper[4744]: I1008 09:23:34.194106 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-8rkml" podStartSLOduration=2.1940775869999998 podStartE2EDuration="2.194077587s" podCreationTimestamp="2025-10-08 09:23:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:23:34.191560414 +0000 UTC m=+709.439205693" watchObservedRunningTime="2025-10-08 09:23:34.194077587 +0000 UTC m=+709.441722836" Oct 08 09:23:34 crc kubenswrapper[4744]: I1008 09:23:34.360429 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-memberlist\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:34 crc kubenswrapper[4744]: I1008 09:23:34.369829 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/349b3e3f-87d0-4e3d-b2dc-c96452a7199c-memberlist\") pod \"speaker-89qxn\" (UID: \"349b3e3f-87d0-4e3d-b2dc-c96452a7199c\") " pod="metallb-system/speaker-89qxn" Oct 08 09:23:34 crc kubenswrapper[4744]: I1008 09:23:34.518190 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-89qxn" Oct 08 09:23:35 crc kubenswrapper[4744]: I1008 09:23:35.202245 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-89qxn" event={"ID":"349b3e3f-87d0-4e3d-b2dc-c96452a7199c","Type":"ContainerStarted","Data":"9b890278e1326eab3e08bb6b76008661dd75ca911f6a36c44778a16cf017ee20"} Oct 08 09:23:35 crc kubenswrapper[4744]: I1008 09:23:35.203758 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-89qxn" event={"ID":"349b3e3f-87d0-4e3d-b2dc-c96452a7199c","Type":"ContainerStarted","Data":"8d5535e92c636de2eb2333af9420be9fcd69286ccf433389ad7afc2f7a8c63cb"} Oct 08 09:23:36 crc kubenswrapper[4744]: I1008 09:23:36.211765 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-89qxn" event={"ID":"349b3e3f-87d0-4e3d-b2dc-c96452a7199c","Type":"ContainerStarted","Data":"f4711c7dfe9499d9f0e48426751ac4bdd9fcb7a18ce37f43271d1a5fb56d4a4a"} Oct 08 09:23:36 crc kubenswrapper[4744]: I1008 09:23:36.214434 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-89qxn" Oct 08 09:23:36 crc kubenswrapper[4744]: I1008 09:23:36.248069 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-89qxn" podStartSLOduration=4.248043559 podStartE2EDuration="4.248043559s" podCreationTimestamp="2025-10-08 09:23:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:23:36.24285624 +0000 UTC m=+711.490501489" watchObservedRunningTime="2025-10-08 09:23:36.248043559 +0000 UTC m=+711.495688788" Oct 08 09:23:42 crc kubenswrapper[4744]: I1008 09:23:42.261751 4744 generic.go:334] "Generic (PLEG): container finished" podID="38b9b97b-50ef-48c1-b917-a6c2f2be304b" containerID="68ef3282f4b58da6395e65e2e4bfab500963723bfd73210085ac046c06ffd414" exitCode=0 Oct 08 09:23:42 crc kubenswrapper[4744]: I1008 09:23:42.261838 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5skh4" event={"ID":"38b9b97b-50ef-48c1-b917-a6c2f2be304b","Type":"ContainerDied","Data":"68ef3282f4b58da6395e65e2e4bfab500963723bfd73210085ac046c06ffd414"} Oct 08 09:23:42 crc kubenswrapper[4744]: I1008 09:23:42.265388 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" event={"ID":"cf7d04b2-1136-4427-8939-e2974337725c","Type":"ContainerStarted","Data":"440e6693ef294d2eb92c806dc7ff633123170665777abc754ac870115f6de871"} Oct 08 09:23:42 crc kubenswrapper[4744]: I1008 09:23:42.265510 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" Oct 08 09:23:42 crc kubenswrapper[4744]: I1008 09:23:42.329650 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" podStartSLOduration=2.021419131 podStartE2EDuration="10.329629477s" podCreationTimestamp="2025-10-08 09:23:32 +0000 UTC" firstStartedPulling="2025-10-08 09:23:33.761613086 +0000 UTC m=+709.009258325" lastFinishedPulling="2025-10-08 09:23:42.069823432 +0000 UTC m=+717.317468671" observedRunningTime="2025-10-08 09:23:42.325809337 +0000 UTC m=+717.573454606" watchObservedRunningTime="2025-10-08 09:23:42.329629477 +0000 UTC m=+717.577274716" Oct 08 09:23:43 crc kubenswrapper[4744]: I1008 09:23:43.037066 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-8rkml" Oct 08 09:23:43 crc kubenswrapper[4744]: I1008 09:23:43.273852 4744 generic.go:334] "Generic (PLEG): container finished" podID="38b9b97b-50ef-48c1-b917-a6c2f2be304b" containerID="f5338c97861b3652795a209a1a7dfe8441a143c6d117588258604cb075b61dc8" exitCode=0 Oct 08 09:23:43 crc kubenswrapper[4744]: I1008 09:23:43.273934 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5skh4" event={"ID":"38b9b97b-50ef-48c1-b917-a6c2f2be304b","Type":"ContainerDied","Data":"f5338c97861b3652795a209a1a7dfe8441a143c6d117588258604cb075b61dc8"} Oct 08 09:23:44 crc kubenswrapper[4744]: I1008 09:23:44.284314 4744 generic.go:334] "Generic (PLEG): container finished" podID="38b9b97b-50ef-48c1-b917-a6c2f2be304b" containerID="96976f5b64ad998ccbd6b70a3f1d308e73c3a5dc47fea6ad242d1e82efb10551" exitCode=0 Oct 08 09:23:44 crc kubenswrapper[4744]: I1008 09:23:44.284402 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5skh4" event={"ID":"38b9b97b-50ef-48c1-b917-a6c2f2be304b","Type":"ContainerDied","Data":"96976f5b64ad998ccbd6b70a3f1d308e73c3a5dc47fea6ad242d1e82efb10551"} Oct 08 09:23:44 crc kubenswrapper[4744]: I1008 09:23:44.526433 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-89qxn" Oct 08 09:23:45 crc kubenswrapper[4744]: I1008 09:23:45.307710 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5skh4" event={"ID":"38b9b97b-50ef-48c1-b917-a6c2f2be304b","Type":"ContainerStarted","Data":"cee93364653392199116e462055517a3681d58a05beff7742bf5b3db451eea6b"} Oct 08 09:23:45 crc kubenswrapper[4744]: I1008 09:23:45.308137 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5skh4" event={"ID":"38b9b97b-50ef-48c1-b917-a6c2f2be304b","Type":"ContainerStarted","Data":"8a55670a533a3ea48f17a1111293be9c24c01c2dc34433b2fb1a141f6316b024"} Oct 08 09:23:45 crc kubenswrapper[4744]: I1008 09:23:45.308164 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5skh4" event={"ID":"38b9b97b-50ef-48c1-b917-a6c2f2be304b","Type":"ContainerStarted","Data":"5bf77daba7df9ef03e50f56762987cfc4d09952798c8d3958b7db97979917a2f"} Oct 08 09:23:45 crc kubenswrapper[4744]: I1008 09:23:45.308182 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5skh4" event={"ID":"38b9b97b-50ef-48c1-b917-a6c2f2be304b","Type":"ContainerStarted","Data":"a86b53c71b15f554158b52b772b741ba2dfe0adf074b8c5bcae322b577114039"} Oct 08 09:23:45 crc kubenswrapper[4744]: I1008 09:23:45.308199 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5skh4" event={"ID":"38b9b97b-50ef-48c1-b917-a6c2f2be304b","Type":"ContainerStarted","Data":"9183c822e2d0fbf516658937b088979378682cb5da203dddcac12133542dba90"} Oct 08 09:23:46 crc kubenswrapper[4744]: I1008 09:23:46.325316 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5skh4" event={"ID":"38b9b97b-50ef-48c1-b917-a6c2f2be304b","Type":"ContainerStarted","Data":"52cc756a69fbc714866488727879675ed788465b022921471f27b46789a08269"} Oct 08 09:23:46 crc kubenswrapper[4744]: I1008 09:23:46.325802 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:46 crc kubenswrapper[4744]: I1008 09:23:46.351147 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-5skh4" podStartSLOduration=5.865191082 podStartE2EDuration="14.351113767s" podCreationTimestamp="2025-10-08 09:23:32 +0000 UTC" firstStartedPulling="2025-10-08 09:23:33.602572621 +0000 UTC m=+708.850217860" lastFinishedPulling="2025-10-08 09:23:42.088495306 +0000 UTC m=+717.336140545" observedRunningTime="2025-10-08 09:23:46.346099474 +0000 UTC m=+721.593744723" watchObservedRunningTime="2025-10-08 09:23:46.351113767 +0000 UTC m=+721.598759006" Oct 08 09:23:47 crc kubenswrapper[4744]: I1008 09:23:47.672189 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-q65z5"] Oct 08 09:23:47 crc kubenswrapper[4744]: I1008 09:23:47.672943 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q65z5" Oct 08 09:23:47 crc kubenswrapper[4744]: I1008 09:23:47.679245 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-hqtzq" Oct 08 09:23:47 crc kubenswrapper[4744]: I1008 09:23:47.679261 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 08 09:23:47 crc kubenswrapper[4744]: I1008 09:23:47.680218 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 08 09:23:47 crc kubenswrapper[4744]: I1008 09:23:47.696075 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5qp4\" (UniqueName: \"kubernetes.io/projected/1b3fd018-8d6c-489f-bbc3-66553cc965d6-kube-api-access-l5qp4\") pod \"openstack-operator-index-q65z5\" (UID: \"1b3fd018-8d6c-489f-bbc3-66553cc965d6\") " pod="openstack-operators/openstack-operator-index-q65z5" Oct 08 09:23:47 crc kubenswrapper[4744]: I1008 09:23:47.704961 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-q65z5"] Oct 08 09:23:47 crc kubenswrapper[4744]: I1008 09:23:47.798073 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5qp4\" (UniqueName: \"kubernetes.io/projected/1b3fd018-8d6c-489f-bbc3-66553cc965d6-kube-api-access-l5qp4\") pod \"openstack-operator-index-q65z5\" (UID: \"1b3fd018-8d6c-489f-bbc3-66553cc965d6\") " pod="openstack-operators/openstack-operator-index-q65z5" Oct 08 09:23:47 crc kubenswrapper[4744]: I1008 09:23:47.854734 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5qp4\" (UniqueName: \"kubernetes.io/projected/1b3fd018-8d6c-489f-bbc3-66553cc965d6-kube-api-access-l5qp4\") pod \"openstack-operator-index-q65z5\" (UID: \"1b3fd018-8d6c-489f-bbc3-66553cc965d6\") " pod="openstack-operators/openstack-operator-index-q65z5" Oct 08 09:23:47 crc kubenswrapper[4744]: I1008 09:23:47.991073 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q65z5" Oct 08 09:23:48 crc kubenswrapper[4744]: I1008 09:23:48.325488 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-q65z5"] Oct 08 09:23:48 crc kubenswrapper[4744]: W1008 09:23:48.340602 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b3fd018_8d6c_489f_bbc3_66553cc965d6.slice/crio-005727462ced24573c7ae90373465c8a46cf8bb295493abe40fedfc9a2786104 WatchSource:0}: Error finding container 005727462ced24573c7ae90373465c8a46cf8bb295493abe40fedfc9a2786104: Status 404 returned error can't find the container with id 005727462ced24573c7ae90373465c8a46cf8bb295493abe40fedfc9a2786104 Oct 08 09:23:48 crc kubenswrapper[4744]: I1008 09:23:48.351063 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q65z5" event={"ID":"1b3fd018-8d6c-489f-bbc3-66553cc965d6","Type":"ContainerStarted","Data":"005727462ced24573c7ae90373465c8a46cf8bb295493abe40fedfc9a2786104"} Oct 08 09:23:48 crc kubenswrapper[4744]: I1008 09:23:48.483116 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:48 crc kubenswrapper[4744]: I1008 09:23:48.551125 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-5skh4" Oct 08 09:23:49 crc kubenswrapper[4744]: I1008 09:23:49.691066 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:23:49 crc kubenswrapper[4744]: I1008 09:23:49.691818 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:23:50 crc kubenswrapper[4744]: I1008 09:23:50.374855 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q65z5" event={"ID":"1b3fd018-8d6c-489f-bbc3-66553cc965d6","Type":"ContainerStarted","Data":"291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5"} Oct 08 09:23:50 crc kubenswrapper[4744]: I1008 09:23:50.407916 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-q65z5" podStartSLOduration=2.495271639 podStartE2EDuration="3.407883517s" podCreationTimestamp="2025-10-08 09:23:47 +0000 UTC" firstStartedPulling="2025-10-08 09:23:48.342664823 +0000 UTC m=+723.590310062" lastFinishedPulling="2025-10-08 09:23:49.255276671 +0000 UTC m=+724.502921940" observedRunningTime="2025-10-08 09:23:50.401656139 +0000 UTC m=+725.649301458" watchObservedRunningTime="2025-10-08 09:23:50.407883517 +0000 UTC m=+725.655528756" Oct 08 09:23:50 crc kubenswrapper[4744]: I1008 09:23:50.836278 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-q65z5"] Oct 08 09:23:51 crc kubenswrapper[4744]: I1008 09:23:51.450515 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-tc4hq"] Oct 08 09:23:51 crc kubenswrapper[4744]: I1008 09:23:51.451746 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tc4hq" Oct 08 09:23:51 crc kubenswrapper[4744]: I1008 09:23:51.472865 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tc4hq"] Oct 08 09:23:51 crc kubenswrapper[4744]: I1008 09:23:51.579028 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtqrk\" (UniqueName: \"kubernetes.io/projected/7f7736f3-f7de-4ff0-ae46-f01bf32693f5-kube-api-access-mtqrk\") pod \"openstack-operator-index-tc4hq\" (UID: \"7f7736f3-f7de-4ff0-ae46-f01bf32693f5\") " pod="openstack-operators/openstack-operator-index-tc4hq" Oct 08 09:23:51 crc kubenswrapper[4744]: I1008 09:23:51.680845 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtqrk\" (UniqueName: \"kubernetes.io/projected/7f7736f3-f7de-4ff0-ae46-f01bf32693f5-kube-api-access-mtqrk\") pod \"openstack-operator-index-tc4hq\" (UID: \"7f7736f3-f7de-4ff0-ae46-f01bf32693f5\") " pod="openstack-operators/openstack-operator-index-tc4hq" Oct 08 09:23:51 crc kubenswrapper[4744]: I1008 09:23:51.715778 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtqrk\" (UniqueName: \"kubernetes.io/projected/7f7736f3-f7de-4ff0-ae46-f01bf32693f5-kube-api-access-mtqrk\") pod \"openstack-operator-index-tc4hq\" (UID: \"7f7736f3-f7de-4ff0-ae46-f01bf32693f5\") " pod="openstack-operators/openstack-operator-index-tc4hq" Oct 08 09:23:51 crc kubenswrapper[4744]: I1008 09:23:51.780605 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tc4hq" Oct 08 09:23:52 crc kubenswrapper[4744]: I1008 09:23:52.007020 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tc4hq"] Oct 08 09:23:52 crc kubenswrapper[4744]: I1008 09:23:52.391946 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-q65z5" podUID="1b3fd018-8d6c-489f-bbc3-66553cc965d6" containerName="registry-server" containerID="cri-o://291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5" gracePeriod=2 Oct 08 09:23:52 crc kubenswrapper[4744]: I1008 09:23:52.392512 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tc4hq" event={"ID":"7f7736f3-f7de-4ff0-ae46-f01bf32693f5","Type":"ContainerStarted","Data":"441109bc1186cd95d81bd955212ae9e48a4bdbf9949d4a7b6f2dadafc46caaea"} Oct 08 09:23:52 crc kubenswrapper[4744]: I1008 09:23:52.778068 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q65z5" Oct 08 09:23:52 crc kubenswrapper[4744]: I1008 09:23:52.799136 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5qp4\" (UniqueName: \"kubernetes.io/projected/1b3fd018-8d6c-489f-bbc3-66553cc965d6-kube-api-access-l5qp4\") pod \"1b3fd018-8d6c-489f-bbc3-66553cc965d6\" (UID: \"1b3fd018-8d6c-489f-bbc3-66553cc965d6\") " Oct 08 09:23:52 crc kubenswrapper[4744]: I1008 09:23:52.808183 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b3fd018-8d6c-489f-bbc3-66553cc965d6-kube-api-access-l5qp4" (OuterVolumeSpecName: "kube-api-access-l5qp4") pod "1b3fd018-8d6c-489f-bbc3-66553cc965d6" (UID: "1b3fd018-8d6c-489f-bbc3-66553cc965d6"). InnerVolumeSpecName "kube-api-access-l5qp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:23:52 crc kubenswrapper[4744]: I1008 09:23:52.902629 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5qp4\" (UniqueName: \"kubernetes.io/projected/1b3fd018-8d6c-489f-bbc3-66553cc965d6-kube-api-access-l5qp4\") on node \"crc\" DevicePath \"\"" Oct 08 09:23:53 crc kubenswrapper[4744]: I1008 09:23:53.403936 4744 generic.go:334] "Generic (PLEG): container finished" podID="1b3fd018-8d6c-489f-bbc3-66553cc965d6" containerID="291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5" exitCode=0 Oct 08 09:23:53 crc kubenswrapper[4744]: I1008 09:23:53.403982 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q65z5" event={"ID":"1b3fd018-8d6c-489f-bbc3-66553cc965d6","Type":"ContainerDied","Data":"291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5"} Oct 08 09:23:53 crc kubenswrapper[4744]: I1008 09:23:53.404024 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-q65z5" event={"ID":"1b3fd018-8d6c-489f-bbc3-66553cc965d6","Type":"ContainerDied","Data":"005727462ced24573c7ae90373465c8a46cf8bb295493abe40fedfc9a2786104"} Oct 08 09:23:53 crc kubenswrapper[4744]: I1008 09:23:53.404044 4744 scope.go:117] "RemoveContainer" containerID="291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5" Oct 08 09:23:53 crc kubenswrapper[4744]: I1008 09:23:53.404087 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-q65z5" Oct 08 09:23:53 crc kubenswrapper[4744]: I1008 09:23:53.406812 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tc4hq" event={"ID":"7f7736f3-f7de-4ff0-ae46-f01bf32693f5","Type":"ContainerStarted","Data":"f3088c538ab6f28d408129ee671bf51159fb19c606246a5868a50f5dc5d8cedc"} Oct 08 09:23:53 crc kubenswrapper[4744]: I1008 09:23:53.427633 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-tc4hq" podStartSLOduration=1.8703995020000002 podStartE2EDuration="2.427608059s" podCreationTimestamp="2025-10-08 09:23:51 +0000 UTC" firstStartedPulling="2025-10-08 09:23:52.018226694 +0000 UTC m=+727.265871933" lastFinishedPulling="2025-10-08 09:23:52.575435251 +0000 UTC m=+727.823080490" observedRunningTime="2025-10-08 09:23:53.426844307 +0000 UTC m=+728.674489576" watchObservedRunningTime="2025-10-08 09:23:53.427608059 +0000 UTC m=+728.675253298" Oct 08 09:23:53 crc kubenswrapper[4744]: I1008 09:23:53.439725 4744 scope.go:117] "RemoveContainer" containerID="291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5" Oct 08 09:23:53 crc kubenswrapper[4744]: E1008 09:23:53.440346 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5\": container with ID starting with 291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5 not found: ID does not exist" containerID="291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5" Oct 08 09:23:53 crc kubenswrapper[4744]: I1008 09:23:53.440426 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5"} err="failed to get container status \"291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5\": rpc error: code = NotFound desc = could not find container \"291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5\": container with ID starting with 291efa81e96efd211c89881e98594a060e96d2a8c5f6e834fa2b0dda107814c5 not found: ID does not exist" Oct 08 09:23:53 crc kubenswrapper[4744]: I1008 09:23:53.467426 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-q65z5"] Oct 08 09:23:53 crc kubenswrapper[4744]: I1008 09:23:53.470274 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-q65z5"] Oct 08 09:23:53 crc kubenswrapper[4744]: I1008 09:23:53.540079 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptt9h" Oct 08 09:23:55 crc kubenswrapper[4744]: I1008 09:23:55.471536 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b3fd018-8d6c-489f-bbc3-66553cc965d6" path="/var/lib/kubelet/pods/1b3fd018-8d6c-489f-bbc3-66553cc965d6/volumes" Oct 08 09:24:01 crc kubenswrapper[4744]: I1008 09:24:01.782017 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-tc4hq" Oct 08 09:24:01 crc kubenswrapper[4744]: I1008 09:24:01.782630 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-tc4hq" Oct 08 09:24:01 crc kubenswrapper[4744]: I1008 09:24:01.814972 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-tc4hq" Oct 08 09:24:02 crc kubenswrapper[4744]: I1008 09:24:02.488896 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-tc4hq" Oct 08 09:24:03 crc kubenswrapper[4744]: I1008 09:24:03.486359 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-5skh4" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.682235 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw"] Oct 08 09:24:09 crc kubenswrapper[4744]: E1008 09:24:09.682781 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b3fd018-8d6c-489f-bbc3-66553cc965d6" containerName="registry-server" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.682795 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b3fd018-8d6c-489f-bbc3-66553cc965d6" containerName="registry-server" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.682911 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b3fd018-8d6c-489f-bbc3-66553cc965d6" containerName="registry-server" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.683706 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.686590 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-pqfjh" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.738406 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw"] Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.766503 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db93bdf5-bdc7-44ba-b635-83ae935f947b-bundle\") pod \"eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw\" (UID: \"db93bdf5-bdc7-44ba-b635-83ae935f947b\") " pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.766761 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5jkw\" (UniqueName: \"kubernetes.io/projected/db93bdf5-bdc7-44ba-b635-83ae935f947b-kube-api-access-p5jkw\") pod \"eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw\" (UID: \"db93bdf5-bdc7-44ba-b635-83ae935f947b\") " pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.766898 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db93bdf5-bdc7-44ba-b635-83ae935f947b-util\") pod \"eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw\" (UID: \"db93bdf5-bdc7-44ba-b635-83ae935f947b\") " pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.868804 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5jkw\" (UniqueName: \"kubernetes.io/projected/db93bdf5-bdc7-44ba-b635-83ae935f947b-kube-api-access-p5jkw\") pod \"eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw\" (UID: \"db93bdf5-bdc7-44ba-b635-83ae935f947b\") " pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.869791 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db93bdf5-bdc7-44ba-b635-83ae935f947b-util\") pod \"eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw\" (UID: \"db93bdf5-bdc7-44ba-b635-83ae935f947b\") " pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.870082 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db93bdf5-bdc7-44ba-b635-83ae935f947b-bundle\") pod \"eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw\" (UID: \"db93bdf5-bdc7-44ba-b635-83ae935f947b\") " pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.870805 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db93bdf5-bdc7-44ba-b635-83ae935f947b-util\") pod \"eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw\" (UID: \"db93bdf5-bdc7-44ba-b635-83ae935f947b\") " pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.870833 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db93bdf5-bdc7-44ba-b635-83ae935f947b-bundle\") pod \"eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw\" (UID: \"db93bdf5-bdc7-44ba-b635-83ae935f947b\") " pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:09 crc kubenswrapper[4744]: I1008 09:24:09.906572 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5jkw\" (UniqueName: \"kubernetes.io/projected/db93bdf5-bdc7-44ba-b635-83ae935f947b-kube-api-access-p5jkw\") pod \"eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw\" (UID: \"db93bdf5-bdc7-44ba-b635-83ae935f947b\") " pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:10 crc kubenswrapper[4744]: I1008 09:24:09.999874 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:10 crc kubenswrapper[4744]: I1008 09:24:10.466133 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw"] Oct 08 09:24:10 crc kubenswrapper[4744]: I1008 09:24:10.524203 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" event={"ID":"db93bdf5-bdc7-44ba-b635-83ae935f947b","Type":"ContainerStarted","Data":"c376f5586abaeb4446ff125fac0807e48c79c454ae0210505b0301f4b4ffe30a"} Oct 08 09:24:11 crc kubenswrapper[4744]: I1008 09:24:11.535445 4744 generic.go:334] "Generic (PLEG): container finished" podID="db93bdf5-bdc7-44ba-b635-83ae935f947b" containerID="2d531ef79a3232ec4159a3377c9c3d22f1cc7adfa19e51048d66a8e240ed5af8" exitCode=0 Oct 08 09:24:11 crc kubenswrapper[4744]: I1008 09:24:11.535521 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" event={"ID":"db93bdf5-bdc7-44ba-b635-83ae935f947b","Type":"ContainerDied","Data":"2d531ef79a3232ec4159a3377c9c3d22f1cc7adfa19e51048d66a8e240ed5af8"} Oct 08 09:24:12 crc kubenswrapper[4744]: I1008 09:24:12.545639 4744 generic.go:334] "Generic (PLEG): container finished" podID="db93bdf5-bdc7-44ba-b635-83ae935f947b" containerID="f5901b653452fc070daab07fade62c7514d1846e1996f626b70c800df6888024" exitCode=0 Oct 08 09:24:12 crc kubenswrapper[4744]: I1008 09:24:12.545689 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" event={"ID":"db93bdf5-bdc7-44ba-b635-83ae935f947b","Type":"ContainerDied","Data":"f5901b653452fc070daab07fade62c7514d1846e1996f626b70c800df6888024"} Oct 08 09:24:13 crc kubenswrapper[4744]: I1008 09:24:13.557040 4744 generic.go:334] "Generic (PLEG): container finished" podID="db93bdf5-bdc7-44ba-b635-83ae935f947b" containerID="d769a73cd50ea328c76c021a6ae6463d626d24637150f4a4900e148779adf532" exitCode=0 Oct 08 09:24:13 crc kubenswrapper[4744]: I1008 09:24:13.557094 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" event={"ID":"db93bdf5-bdc7-44ba-b635-83ae935f947b","Type":"ContainerDied","Data":"d769a73cd50ea328c76c021a6ae6463d626d24637150f4a4900e148779adf532"} Oct 08 09:24:14 crc kubenswrapper[4744]: I1008 09:24:14.837919 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:14 crc kubenswrapper[4744]: I1008 09:24:14.952453 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db93bdf5-bdc7-44ba-b635-83ae935f947b-util\") pod \"db93bdf5-bdc7-44ba-b635-83ae935f947b\" (UID: \"db93bdf5-bdc7-44ba-b635-83ae935f947b\") " Oct 08 09:24:14 crc kubenswrapper[4744]: I1008 09:24:14.952586 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5jkw\" (UniqueName: \"kubernetes.io/projected/db93bdf5-bdc7-44ba-b635-83ae935f947b-kube-api-access-p5jkw\") pod \"db93bdf5-bdc7-44ba-b635-83ae935f947b\" (UID: \"db93bdf5-bdc7-44ba-b635-83ae935f947b\") " Oct 08 09:24:14 crc kubenswrapper[4744]: I1008 09:24:14.952618 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db93bdf5-bdc7-44ba-b635-83ae935f947b-bundle\") pod \"db93bdf5-bdc7-44ba-b635-83ae935f947b\" (UID: \"db93bdf5-bdc7-44ba-b635-83ae935f947b\") " Oct 08 09:24:14 crc kubenswrapper[4744]: I1008 09:24:14.953773 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db93bdf5-bdc7-44ba-b635-83ae935f947b-bundle" (OuterVolumeSpecName: "bundle") pod "db93bdf5-bdc7-44ba-b635-83ae935f947b" (UID: "db93bdf5-bdc7-44ba-b635-83ae935f947b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:24:14 crc kubenswrapper[4744]: I1008 09:24:14.961558 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db93bdf5-bdc7-44ba-b635-83ae935f947b-kube-api-access-p5jkw" (OuterVolumeSpecName: "kube-api-access-p5jkw") pod "db93bdf5-bdc7-44ba-b635-83ae935f947b" (UID: "db93bdf5-bdc7-44ba-b635-83ae935f947b"). InnerVolumeSpecName "kube-api-access-p5jkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:24:14 crc kubenswrapper[4744]: I1008 09:24:14.976870 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db93bdf5-bdc7-44ba-b635-83ae935f947b-util" (OuterVolumeSpecName: "util") pod "db93bdf5-bdc7-44ba-b635-83ae935f947b" (UID: "db93bdf5-bdc7-44ba-b635-83ae935f947b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:24:15 crc kubenswrapper[4744]: I1008 09:24:15.054986 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5jkw\" (UniqueName: \"kubernetes.io/projected/db93bdf5-bdc7-44ba-b635-83ae935f947b-kube-api-access-p5jkw\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:15 crc kubenswrapper[4744]: I1008 09:24:15.055042 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/db93bdf5-bdc7-44ba-b635-83ae935f947b-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:15 crc kubenswrapper[4744]: I1008 09:24:15.055052 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/db93bdf5-bdc7-44ba-b635-83ae935f947b-util\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:15 crc kubenswrapper[4744]: I1008 09:24:15.574384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" event={"ID":"db93bdf5-bdc7-44ba-b635-83ae935f947b","Type":"ContainerDied","Data":"c376f5586abaeb4446ff125fac0807e48c79c454ae0210505b0301f4b4ffe30a"} Oct 08 09:24:15 crc kubenswrapper[4744]: I1008 09:24:15.574471 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c376f5586abaeb4446ff125fac0807e48c79c454ae0210505b0301f4b4ffe30a" Oct 08 09:24:15 crc kubenswrapper[4744]: I1008 09:24:15.575271 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.134158 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8h78h"] Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.134415 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" podUID="216b2020-4266-4895-85cb-f0939118b7a8" containerName="controller-manager" containerID="cri-o://cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa" gracePeriod=30 Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.231050 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n"] Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.231403 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" podUID="ff9cb19e-ad39-4636-b302-57b5f39b0072" containerName="route-controller-manager" containerID="cri-o://4c4a65acb3b3126de5a564faad9818df0a8a4622eddd0e6b71cda52923e9075c" gracePeriod=30 Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.608765 4744 generic.go:334] "Generic (PLEG): container finished" podID="ff9cb19e-ad39-4636-b302-57b5f39b0072" containerID="4c4a65acb3b3126de5a564faad9818df0a8a4622eddd0e6b71cda52923e9075c" exitCode=0 Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.608923 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" event={"ID":"ff9cb19e-ad39-4636-b302-57b5f39b0072","Type":"ContainerDied","Data":"4c4a65acb3b3126de5a564faad9818df0a8a4622eddd0e6b71cda52923e9075c"} Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.609806 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.610973 4744 generic.go:334] "Generic (PLEG): container finished" podID="216b2020-4266-4895-85cb-f0939118b7a8" containerID="cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa" exitCode=0 Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.611008 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" event={"ID":"216b2020-4266-4895-85cb-f0939118b7a8","Type":"ContainerDied","Data":"cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa"} Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.611032 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" event={"ID":"216b2020-4266-4895-85cb-f0939118b7a8","Type":"ContainerDied","Data":"1b48011b9277f1094d7e1f74c108f758e6656dec57b2fdc20c175d4df222475c"} Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.611050 4744 scope.go:117] "RemoveContainer" containerID="cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.635665 4744 scope.go:117] "RemoveContainer" containerID="cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa" Oct 08 09:24:16 crc kubenswrapper[4744]: E1008 09:24:16.636350 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa\": container with ID starting with cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa not found: ID does not exist" containerID="cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.636473 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa"} err="failed to get container status \"cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa\": rpc error: code = NotFound desc = could not find container \"cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa\": container with ID starting with cfa5324dab7161e081c665de9f33ad51a84a5bffc2fb3d20fbbd383683b517aa not found: ID does not exist" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.700917 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.783442 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-client-ca\") pod \"216b2020-4266-4895-85cb-f0939118b7a8\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.783527 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-config\") pod \"216b2020-4266-4895-85cb-f0939118b7a8\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.783577 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-proxy-ca-bundles\") pod \"216b2020-4266-4895-85cb-f0939118b7a8\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.783662 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/216b2020-4266-4895-85cb-f0939118b7a8-serving-cert\") pod \"216b2020-4266-4895-85cb-f0939118b7a8\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.784533 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "216b2020-4266-4895-85cb-f0939118b7a8" (UID: "216b2020-4266-4895-85cb-f0939118b7a8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.784601 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-config" (OuterVolumeSpecName: "config") pod "216b2020-4266-4895-85cb-f0939118b7a8" (UID: "216b2020-4266-4895-85cb-f0939118b7a8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.784649 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlf55\" (UniqueName: \"kubernetes.io/projected/216b2020-4266-4895-85cb-f0939118b7a8-kube-api-access-mlf55\") pod \"216b2020-4266-4895-85cb-f0939118b7a8\" (UID: \"216b2020-4266-4895-85cb-f0939118b7a8\") " Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.784850 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-client-ca" (OuterVolumeSpecName: "client-ca") pod "216b2020-4266-4895-85cb-f0939118b7a8" (UID: "216b2020-4266-4895-85cb-f0939118b7a8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.785195 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.785219 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.785292 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/216b2020-4266-4895-85cb-f0939118b7a8-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.791331 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/216b2020-4266-4895-85cb-f0939118b7a8-kube-api-access-mlf55" (OuterVolumeSpecName: "kube-api-access-mlf55") pod "216b2020-4266-4895-85cb-f0939118b7a8" (UID: "216b2020-4266-4895-85cb-f0939118b7a8"). InnerVolumeSpecName "kube-api-access-mlf55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.791347 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/216b2020-4266-4895-85cb-f0939118b7a8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "216b2020-4266-4895-85cb-f0939118b7a8" (UID: "216b2020-4266-4895-85cb-f0939118b7a8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.886043 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff9cb19e-ad39-4636-b302-57b5f39b0072-client-ca\") pod \"ff9cb19e-ad39-4636-b302-57b5f39b0072\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.886650 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff9cb19e-ad39-4636-b302-57b5f39b0072-serving-cert\") pod \"ff9cb19e-ad39-4636-b302-57b5f39b0072\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.886682 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff9cb19e-ad39-4636-b302-57b5f39b0072-config\") pod \"ff9cb19e-ad39-4636-b302-57b5f39b0072\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.886702 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svjqj\" (UniqueName: \"kubernetes.io/projected/ff9cb19e-ad39-4636-b302-57b5f39b0072-kube-api-access-svjqj\") pod \"ff9cb19e-ad39-4636-b302-57b5f39b0072\" (UID: \"ff9cb19e-ad39-4636-b302-57b5f39b0072\") " Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.886945 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/216b2020-4266-4895-85cb-f0939118b7a8-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.886958 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlf55\" (UniqueName: \"kubernetes.io/projected/216b2020-4266-4895-85cb-f0939118b7a8-kube-api-access-mlf55\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.887118 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff9cb19e-ad39-4636-b302-57b5f39b0072-client-ca" (OuterVolumeSpecName: "client-ca") pod "ff9cb19e-ad39-4636-b302-57b5f39b0072" (UID: "ff9cb19e-ad39-4636-b302-57b5f39b0072"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.887205 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff9cb19e-ad39-4636-b302-57b5f39b0072-config" (OuterVolumeSpecName: "config") pod "ff9cb19e-ad39-4636-b302-57b5f39b0072" (UID: "ff9cb19e-ad39-4636-b302-57b5f39b0072"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.890872 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff9cb19e-ad39-4636-b302-57b5f39b0072-kube-api-access-svjqj" (OuterVolumeSpecName: "kube-api-access-svjqj") pod "ff9cb19e-ad39-4636-b302-57b5f39b0072" (UID: "ff9cb19e-ad39-4636-b302-57b5f39b0072"). InnerVolumeSpecName "kube-api-access-svjqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.891195 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff9cb19e-ad39-4636-b302-57b5f39b0072-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ff9cb19e-ad39-4636-b302-57b5f39b0072" (UID: "ff9cb19e-ad39-4636-b302-57b5f39b0072"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.987836 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ff9cb19e-ad39-4636-b302-57b5f39b0072-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.987882 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff9cb19e-ad39-4636-b302-57b5f39b0072-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.987896 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff9cb19e-ad39-4636-b302-57b5f39b0072-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:16 crc kubenswrapper[4744]: I1008 09:24:16.987910 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svjqj\" (UniqueName: \"kubernetes.io/projected/ff9cb19e-ad39-4636-b302-57b5f39b0072-kube-api-access-svjqj\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.451644 4744 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-8h78h container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.451735 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" podUID="216b2020-4266-4895-85cb-f0939118b7a8" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.470768 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7"] Oct 08 09:24:17 crc kubenswrapper[4744]: E1008 09:24:17.471420 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff9cb19e-ad39-4636-b302-57b5f39b0072" containerName="route-controller-manager" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.473574 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff9cb19e-ad39-4636-b302-57b5f39b0072" containerName="route-controller-manager" Oct 08 09:24:17 crc kubenswrapper[4744]: E1008 09:24:17.473820 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db93bdf5-bdc7-44ba-b635-83ae935f947b" containerName="extract" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.473905 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="db93bdf5-bdc7-44ba-b635-83ae935f947b" containerName="extract" Oct 08 09:24:17 crc kubenswrapper[4744]: E1008 09:24:17.473980 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db93bdf5-bdc7-44ba-b635-83ae935f947b" containerName="pull" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.474064 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="db93bdf5-bdc7-44ba-b635-83ae935f947b" containerName="pull" Oct 08 09:24:17 crc kubenswrapper[4744]: E1008 09:24:17.474162 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="216b2020-4266-4895-85cb-f0939118b7a8" containerName="controller-manager" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.474234 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="216b2020-4266-4895-85cb-f0939118b7a8" containerName="controller-manager" Oct 08 09:24:17 crc kubenswrapper[4744]: E1008 09:24:17.474321 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db93bdf5-bdc7-44ba-b635-83ae935f947b" containerName="util" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.474416 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="db93bdf5-bdc7-44ba-b635-83ae935f947b" containerName="util" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.474672 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff9cb19e-ad39-4636-b302-57b5f39b0072" containerName="route-controller-manager" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.474778 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="216b2020-4266-4895-85cb-f0939118b7a8" containerName="controller-manager" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.474896 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="db93bdf5-bdc7-44ba-b635-83ae935f947b" containerName="extract" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.475512 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-59c45dfc9-2w5cr"] Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.475739 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.476831 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7"] Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.477102 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.478852 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59c45dfc9-2w5cr"] Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.598437 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74c3c709-b4a8-41c8-acdc-ba423fc250b6-serving-cert\") pod \"route-controller-manager-fc74f9f6-6sdn7\" (UID: \"74c3c709-b4a8-41c8-acdc-ba423fc250b6\") " pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.598519 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6twj5\" (UniqueName: \"kubernetes.io/projected/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-kube-api-access-6twj5\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.598551 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74c3c709-b4a8-41c8-acdc-ba423fc250b6-config\") pod \"route-controller-manager-fc74f9f6-6sdn7\" (UID: \"74c3c709-b4a8-41c8-acdc-ba423fc250b6\") " pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.598581 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-proxy-ca-bundles\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.598621 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/74c3c709-b4a8-41c8-acdc-ba423fc250b6-client-ca\") pod \"route-controller-manager-fc74f9f6-6sdn7\" (UID: \"74c3c709-b4a8-41c8-acdc-ba423fc250b6\") " pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.598674 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f87z8\" (UniqueName: \"kubernetes.io/projected/74c3c709-b4a8-41c8-acdc-ba423fc250b6-kube-api-access-f87z8\") pod \"route-controller-manager-fc74f9f6-6sdn7\" (UID: \"74c3c709-b4a8-41c8-acdc-ba423fc250b6\") " pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.598693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-config\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.598725 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-client-ca\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.598745 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-serving-cert\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.619433 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" event={"ID":"ff9cb19e-ad39-4636-b302-57b5f39b0072","Type":"ContainerDied","Data":"0de88f35f62efd38ec83f29221423ed7ec071b20a907500a557a2624fe0e69bc"} Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.619699 4744 scope.go:117] "RemoveContainer" containerID="4c4a65acb3b3126de5a564faad9818df0a8a4622eddd0e6b71cda52923e9075c" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.620058 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.621668 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8h78h" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.647314 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8h78h"] Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.656823 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8h78h"] Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.666800 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n"] Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.673701 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-hns6n"] Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.700181 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f87z8\" (UniqueName: \"kubernetes.io/projected/74c3c709-b4a8-41c8-acdc-ba423fc250b6-kube-api-access-f87z8\") pod \"route-controller-manager-fc74f9f6-6sdn7\" (UID: \"74c3c709-b4a8-41c8-acdc-ba423fc250b6\") " pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.700756 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-config\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.700855 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-client-ca\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.701795 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-serving-cert\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.702719 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74c3c709-b4a8-41c8-acdc-ba423fc250b6-serving-cert\") pod \"route-controller-manager-fc74f9f6-6sdn7\" (UID: \"74c3c709-b4a8-41c8-acdc-ba423fc250b6\") " pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.702839 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6twj5\" (UniqueName: \"kubernetes.io/projected/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-kube-api-access-6twj5\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.702933 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74c3c709-b4a8-41c8-acdc-ba423fc250b6-config\") pod \"route-controller-manager-fc74f9f6-6sdn7\" (UID: \"74c3c709-b4a8-41c8-acdc-ba423fc250b6\") " pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.703018 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-proxy-ca-bundles\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.703353 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/74c3c709-b4a8-41c8-acdc-ba423fc250b6-client-ca\") pod \"route-controller-manager-fc74f9f6-6sdn7\" (UID: \"74c3c709-b4a8-41c8-acdc-ba423fc250b6\") " pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.704013 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74c3c709-b4a8-41c8-acdc-ba423fc250b6-config\") pod \"route-controller-manager-fc74f9f6-6sdn7\" (UID: \"74c3c709-b4a8-41c8-acdc-ba423fc250b6\") " pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.701751 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-client-ca\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.704245 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/74c3c709-b4a8-41c8-acdc-ba423fc250b6-client-ca\") pod \"route-controller-manager-fc74f9f6-6sdn7\" (UID: \"74c3c709-b4a8-41c8-acdc-ba423fc250b6\") " pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.714504 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-proxy-ca-bundles\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.715513 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-config\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.717815 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-serving-cert\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.721466 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74c3c709-b4a8-41c8-acdc-ba423fc250b6-serving-cert\") pod \"route-controller-manager-fc74f9f6-6sdn7\" (UID: \"74c3c709-b4a8-41c8-acdc-ba423fc250b6\") " pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.722202 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6twj5\" (UniqueName: \"kubernetes.io/projected/ae88f416-d62c-4b5e-8a83-b9af29ba2a93-kube-api-access-6twj5\") pod \"controller-manager-59c45dfc9-2w5cr\" (UID: \"ae88f416-d62c-4b5e-8a83-b9af29ba2a93\") " pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.726032 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f87z8\" (UniqueName: \"kubernetes.io/projected/74c3c709-b4a8-41c8-acdc-ba423fc250b6-kube-api-access-f87z8\") pod \"route-controller-manager-fc74f9f6-6sdn7\" (UID: \"74c3c709-b4a8-41c8-acdc-ba423fc250b6\") " pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.811734 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:17 crc kubenswrapper[4744]: I1008 09:24:17.829206 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:18 crc kubenswrapper[4744]: I1008 09:24:18.078659 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7"] Oct 08 09:24:18 crc kubenswrapper[4744]: I1008 09:24:18.125959 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59c45dfc9-2w5cr"] Oct 08 09:24:18 crc kubenswrapper[4744]: W1008 09:24:18.132346 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae88f416_d62c_4b5e_8a83_b9af29ba2a93.slice/crio-7f3f68e18781768430a9c4f68088b1c15eb30040e04ca9ebb0bb026e09bf020a WatchSource:0}: Error finding container 7f3f68e18781768430a9c4f68088b1c15eb30040e04ca9ebb0bb026e09bf020a: Status 404 returned error can't find the container with id 7f3f68e18781768430a9c4f68088b1c15eb30040e04ca9ebb0bb026e09bf020a Oct 08 09:24:18 crc kubenswrapper[4744]: I1008 09:24:18.629934 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" event={"ID":"74c3c709-b4a8-41c8-acdc-ba423fc250b6","Type":"ContainerStarted","Data":"fcc242befac1a3a00c55a4c543fdba3d7eb8be5c7b208bf0c4ccb1cefd01d08b"} Oct 08 09:24:18 crc kubenswrapper[4744]: I1008 09:24:18.630401 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:18 crc kubenswrapper[4744]: I1008 09:24:18.630416 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" event={"ID":"74c3c709-b4a8-41c8-acdc-ba423fc250b6","Type":"ContainerStarted","Data":"88df18a50480b5f11ac5d5519536933bdbe1157b089c686764d447c3399c56db"} Oct 08 09:24:18 crc kubenswrapper[4744]: I1008 09:24:18.632803 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" event={"ID":"ae88f416-d62c-4b5e-8a83-b9af29ba2a93","Type":"ContainerStarted","Data":"c78f117b3542fcff3777974af75311350d2f6ae585e3862148ba476e4bc8eb25"} Oct 08 09:24:18 crc kubenswrapper[4744]: I1008 09:24:18.632863 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" event={"ID":"ae88f416-d62c-4b5e-8a83-b9af29ba2a93","Type":"ContainerStarted","Data":"7f3f68e18781768430a9c4f68088b1c15eb30040e04ca9ebb0bb026e09bf020a"} Oct 08 09:24:18 crc kubenswrapper[4744]: I1008 09:24:18.633060 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:18 crc kubenswrapper[4744]: I1008 09:24:18.640774 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" Oct 08 09:24:18 crc kubenswrapper[4744]: I1008 09:24:18.678835 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" podStartSLOduration=2.678810942 podStartE2EDuration="2.678810942s" podCreationTimestamp="2025-10-08 09:24:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:24:18.67770344 +0000 UTC m=+753.925348679" watchObservedRunningTime="2025-10-08 09:24:18.678810942 +0000 UTC m=+753.926456181" Oct 08 09:24:18 crc kubenswrapper[4744]: I1008 09:24:18.868334 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-fc74f9f6-6sdn7" Oct 08 09:24:18 crc kubenswrapper[4744]: I1008 09:24:18.894870 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-59c45dfc9-2w5cr" podStartSLOduration=2.894848534 podStartE2EDuration="2.894848534s" podCreationTimestamp="2025-10-08 09:24:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:24:18.698482375 +0000 UTC m=+753.946127634" watchObservedRunningTime="2025-10-08 09:24:18.894848534 +0000 UTC m=+754.142493773" Oct 08 09:24:19 crc kubenswrapper[4744]: I1008 09:24:19.461741 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="216b2020-4266-4895-85cb-f0939118b7a8" path="/var/lib/kubelet/pods/216b2020-4266-4895-85cb-f0939118b7a8/volumes" Oct 08 09:24:19 crc kubenswrapper[4744]: I1008 09:24:19.462481 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff9cb19e-ad39-4636-b302-57b5f39b0072" path="/var/lib/kubelet/pods/ff9cb19e-ad39-4636-b302-57b5f39b0072/volumes" Oct 08 09:24:19 crc kubenswrapper[4744]: I1008 09:24:19.690065 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:24:19 crc kubenswrapper[4744]: I1008 09:24:19.690129 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:24:19 crc kubenswrapper[4744]: I1008 09:24:19.690175 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:24:19 crc kubenswrapper[4744]: I1008 09:24:19.690915 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"12c8d8cd4abdd9a59c0cfbc9852b990e3820f44c3e436a88bb84a00c67314d34"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 09:24:19 crc kubenswrapper[4744]: I1008 09:24:19.690980 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://12c8d8cd4abdd9a59c0cfbc9852b990e3820f44c3e436a88bb84a00c67314d34" gracePeriod=600 Oct 08 09:24:20 crc kubenswrapper[4744]: I1008 09:24:20.653061 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="12c8d8cd4abdd9a59c0cfbc9852b990e3820f44c3e436a88bb84a00c67314d34" exitCode=0 Oct 08 09:24:20 crc kubenswrapper[4744]: I1008 09:24:20.653283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"12c8d8cd4abdd9a59c0cfbc9852b990e3820f44c3e436a88bb84a00c67314d34"} Oct 08 09:24:20 crc kubenswrapper[4744]: I1008 09:24:20.653919 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"06edcaf88cdb63c2778de4ebf46b1539980361ea9f30176a88a8565819ce43a4"} Oct 08 09:24:20 crc kubenswrapper[4744]: I1008 09:24:20.653995 4744 scope.go:117] "RemoveContainer" containerID="8c068a91bb7d0e26249b43fab5fe60ecb8e806b9d40583c0517d3afb38ddf715" Oct 08 09:24:21 crc kubenswrapper[4744]: I1008 09:24:21.872348 4744 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 09:24:22 crc kubenswrapper[4744]: I1008 09:24:22.136564 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr"] Oct 08 09:24:22 crc kubenswrapper[4744]: I1008 09:24:22.137499 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr" Oct 08 09:24:22 crc kubenswrapper[4744]: I1008 09:24:22.143267 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-c8jcx" Oct 08 09:24:22 crc kubenswrapper[4744]: I1008 09:24:22.199599 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr"] Oct 08 09:24:22 crc kubenswrapper[4744]: I1008 09:24:22.292809 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rtnk\" (UniqueName: \"kubernetes.io/projected/b575f955-1e82-4afa-a84b-842c2ba7e47b-kube-api-access-6rtnk\") pod \"openstack-operator-controller-operator-55f65988b-bjktr\" (UID: \"b575f955-1e82-4afa-a84b-842c2ba7e47b\") " pod="openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr" Oct 08 09:24:22 crc kubenswrapper[4744]: I1008 09:24:22.394543 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rtnk\" (UniqueName: \"kubernetes.io/projected/b575f955-1e82-4afa-a84b-842c2ba7e47b-kube-api-access-6rtnk\") pod \"openstack-operator-controller-operator-55f65988b-bjktr\" (UID: \"b575f955-1e82-4afa-a84b-842c2ba7e47b\") " pod="openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr" Oct 08 09:24:22 crc kubenswrapper[4744]: I1008 09:24:22.433311 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rtnk\" (UniqueName: \"kubernetes.io/projected/b575f955-1e82-4afa-a84b-842c2ba7e47b-kube-api-access-6rtnk\") pod \"openstack-operator-controller-operator-55f65988b-bjktr\" (UID: \"b575f955-1e82-4afa-a84b-842c2ba7e47b\") " pod="openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr" Oct 08 09:24:22 crc kubenswrapper[4744]: I1008 09:24:22.456506 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr" Oct 08 09:24:22 crc kubenswrapper[4744]: I1008 09:24:22.952259 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr"] Oct 08 09:24:22 crc kubenswrapper[4744]: W1008 09:24:22.957159 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb575f955_1e82_4afa_a84b_842c2ba7e47b.slice/crio-9c86b2a365ce1aa94d7c7884202e6b9f49b703f945a0f4c19001ddafbcad28e7 WatchSource:0}: Error finding container 9c86b2a365ce1aa94d7c7884202e6b9f49b703f945a0f4c19001ddafbcad28e7: Status 404 returned error can't find the container with id 9c86b2a365ce1aa94d7c7884202e6b9f49b703f945a0f4c19001ddafbcad28e7 Oct 08 09:24:23 crc kubenswrapper[4744]: I1008 09:24:23.688799 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr" event={"ID":"b575f955-1e82-4afa-a84b-842c2ba7e47b","Type":"ContainerStarted","Data":"9c86b2a365ce1aa94d7c7884202e6b9f49b703f945a0f4c19001ddafbcad28e7"} Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.563446 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b7845"] Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.565789 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.603344 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b7845"] Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.659500 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-catalog-content\") pod \"redhat-operators-b7845\" (UID: \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\") " pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.659635 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-utilities\") pod \"redhat-operators-b7845\" (UID: \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\") " pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.661030 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69j8l\" (UniqueName: \"kubernetes.io/projected/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-kube-api-access-69j8l\") pod \"redhat-operators-b7845\" (UID: \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\") " pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.762824 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69j8l\" (UniqueName: \"kubernetes.io/projected/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-kube-api-access-69j8l\") pod \"redhat-operators-b7845\" (UID: \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\") " pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.762928 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-catalog-content\") pod \"redhat-operators-b7845\" (UID: \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\") " pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.762959 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-utilities\") pod \"redhat-operators-b7845\" (UID: \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\") " pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.763824 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-utilities\") pod \"redhat-operators-b7845\" (UID: \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\") " pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.764654 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-catalog-content\") pod \"redhat-operators-b7845\" (UID: \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\") " pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.797148 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69j8l\" (UniqueName: \"kubernetes.io/projected/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-kube-api-access-69j8l\") pod \"redhat-operators-b7845\" (UID: \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\") " pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:24 crc kubenswrapper[4744]: I1008 09:24:24.891590 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:25 crc kubenswrapper[4744]: I1008 09:24:25.541319 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b7845"] Oct 08 09:24:25 crc kubenswrapper[4744]: W1008 09:24:25.561363 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bd561d0_b3a4_42cf_8e1a_ff87abe359b6.slice/crio-6668a5a8fbfa30e4ef5074d3a8faa8110cee2b2cb0f9206c51c9209a8b101132 WatchSource:0}: Error finding container 6668a5a8fbfa30e4ef5074d3a8faa8110cee2b2cb0f9206c51c9209a8b101132: Status 404 returned error can't find the container with id 6668a5a8fbfa30e4ef5074d3a8faa8110cee2b2cb0f9206c51c9209a8b101132 Oct 08 09:24:25 crc kubenswrapper[4744]: I1008 09:24:25.724898 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b7845" event={"ID":"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6","Type":"ContainerStarted","Data":"6668a5a8fbfa30e4ef5074d3a8faa8110cee2b2cb0f9206c51c9209a8b101132"} Oct 08 09:24:26 crc kubenswrapper[4744]: I1008 09:24:26.734253 4744 generic.go:334] "Generic (PLEG): container finished" podID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" containerID="b312b4b496f4889472ec7877fe31f2410ff3ae322cc1b679cc053a5ed13ac344" exitCode=0 Oct 08 09:24:26 crc kubenswrapper[4744]: I1008 09:24:26.734341 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b7845" event={"ID":"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6","Type":"ContainerDied","Data":"b312b4b496f4889472ec7877fe31f2410ff3ae322cc1b679cc053a5ed13ac344"} Oct 08 09:24:29 crc kubenswrapper[4744]: I1008 09:24:29.772010 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b7845" event={"ID":"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6","Type":"ContainerStarted","Data":"d4071b39b3d98698df0f2858ad67c741edafdc72290671dd207031151826ec7c"} Oct 08 09:24:29 crc kubenswrapper[4744]: I1008 09:24:29.773847 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr" event={"ID":"b575f955-1e82-4afa-a84b-842c2ba7e47b","Type":"ContainerStarted","Data":"3034e57aaee6d31f5cad39abd2d45541bb77334c07352a4bcdaa2bd680882480"} Oct 08 09:24:30 crc kubenswrapper[4744]: I1008 09:24:30.782530 4744 generic.go:334] "Generic (PLEG): container finished" podID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" containerID="d4071b39b3d98698df0f2858ad67c741edafdc72290671dd207031151826ec7c" exitCode=0 Oct 08 09:24:30 crc kubenswrapper[4744]: I1008 09:24:30.782641 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b7845" event={"ID":"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6","Type":"ContainerDied","Data":"d4071b39b3d98698df0f2858ad67c741edafdc72290671dd207031151826ec7c"} Oct 08 09:24:32 crc kubenswrapper[4744]: I1008 09:24:32.958218 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9t7n8"] Oct 08 09:24:32 crc kubenswrapper[4744]: I1008 09:24:32.960032 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:32 crc kubenswrapper[4744]: I1008 09:24:32.977212 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9t7n8"] Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.108966 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-utilities\") pod \"community-operators-9t7n8\" (UID: \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\") " pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.109053 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk9nz\" (UniqueName: \"kubernetes.io/projected/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-kube-api-access-lk9nz\") pod \"community-operators-9t7n8\" (UID: \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\") " pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.109086 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-catalog-content\") pod \"community-operators-9t7n8\" (UID: \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\") " pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.210302 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-utilities\") pod \"community-operators-9t7n8\" (UID: \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\") " pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.210375 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk9nz\" (UniqueName: \"kubernetes.io/projected/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-kube-api-access-lk9nz\") pod \"community-operators-9t7n8\" (UID: \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\") " pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.210421 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-catalog-content\") pod \"community-operators-9t7n8\" (UID: \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\") " pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.210878 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-utilities\") pod \"community-operators-9t7n8\" (UID: \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\") " pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.210907 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-catalog-content\") pod \"community-operators-9t7n8\" (UID: \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\") " pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.239487 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk9nz\" (UniqueName: \"kubernetes.io/projected/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-kube-api-access-lk9nz\") pod \"community-operators-9t7n8\" (UID: \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\") " pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.314968 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.826833 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b7845" event={"ID":"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6","Type":"ContainerStarted","Data":"8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98"} Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.834734 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr" event={"ID":"b575f955-1e82-4afa-a84b-842c2ba7e47b","Type":"ContainerStarted","Data":"7677ae72d5d4a7a4801ec2e1b8370c3172ec9eea87e8a6f635892d5c105b1d16"} Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.834978 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.869810 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b7845" podStartSLOduration=5.369812367 podStartE2EDuration="9.869783251s" podCreationTimestamp="2025-10-08 09:24:24 +0000 UTC" firstStartedPulling="2025-10-08 09:24:28.147466944 +0000 UTC m=+763.395112183" lastFinishedPulling="2025-10-08 09:24:32.647437828 +0000 UTC m=+767.895083067" observedRunningTime="2025-10-08 09:24:33.863572304 +0000 UTC m=+769.111217553" watchObservedRunningTime="2025-10-08 09:24:33.869783251 +0000 UTC m=+769.117428490" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.954650 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr" podStartSLOduration=2.191500679 podStartE2EDuration="11.954631259s" podCreationTimestamp="2025-10-08 09:24:22 +0000 UTC" firstStartedPulling="2025-10-08 09:24:22.960135538 +0000 UTC m=+758.207780777" lastFinishedPulling="2025-10-08 09:24:32.723266118 +0000 UTC m=+767.970911357" observedRunningTime="2025-10-08 09:24:33.908489569 +0000 UTC m=+769.156134828" watchObservedRunningTime="2025-10-08 09:24:33.954631259 +0000 UTC m=+769.202276488" Oct 08 09:24:33 crc kubenswrapper[4744]: I1008 09:24:33.955748 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9t7n8"] Oct 08 09:24:34 crc kubenswrapper[4744]: I1008 09:24:34.844966 4744 generic.go:334] "Generic (PLEG): container finished" podID="417b66f9-47a8-4fc7-9d8e-e56c20bc5657" containerID="59cb652825c3d7e2175db99d6c779385f57bf4fffb86b5f0b3dcb115d4b13824" exitCode=0 Oct 08 09:24:34 crc kubenswrapper[4744]: I1008 09:24:34.845078 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t7n8" event={"ID":"417b66f9-47a8-4fc7-9d8e-e56c20bc5657","Type":"ContainerDied","Data":"59cb652825c3d7e2175db99d6c779385f57bf4fffb86b5f0b3dcb115d4b13824"} Oct 08 09:24:34 crc kubenswrapper[4744]: I1008 09:24:34.845327 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t7n8" event={"ID":"417b66f9-47a8-4fc7-9d8e-e56c20bc5657","Type":"ContainerStarted","Data":"d4f4e6f660edebb65beaf3ac10f533f3f5e08903b972b6545a6a05092925ed60"} Oct 08 09:24:34 crc kubenswrapper[4744]: I1008 09:24:34.849571 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-55f65988b-bjktr" Oct 08 09:24:34 crc kubenswrapper[4744]: I1008 09:24:34.892486 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:34 crc kubenswrapper[4744]: I1008 09:24:34.892591 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:35 crc kubenswrapper[4744]: I1008 09:24:35.863718 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t7n8" event={"ID":"417b66f9-47a8-4fc7-9d8e-e56c20bc5657","Type":"ContainerStarted","Data":"a8f3d2b19bde5f9c4939613ebd4646b2bf2ec1bc0b5ddf056ee8a0525fd46820"} Oct 08 09:24:35 crc kubenswrapper[4744]: I1008 09:24:35.946697 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b7845" podUID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" containerName="registry-server" probeResult="failure" output=< Oct 08 09:24:35 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 09:24:35 crc kubenswrapper[4744]: > Oct 08 09:24:36 crc kubenswrapper[4744]: I1008 09:24:36.902240 4744 generic.go:334] "Generic (PLEG): container finished" podID="417b66f9-47a8-4fc7-9d8e-e56c20bc5657" containerID="a8f3d2b19bde5f9c4939613ebd4646b2bf2ec1bc0b5ddf056ee8a0525fd46820" exitCode=0 Oct 08 09:24:36 crc kubenswrapper[4744]: I1008 09:24:36.902362 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t7n8" event={"ID":"417b66f9-47a8-4fc7-9d8e-e56c20bc5657","Type":"ContainerDied","Data":"a8f3d2b19bde5f9c4939613ebd4646b2bf2ec1bc0b5ddf056ee8a0525fd46820"} Oct 08 09:24:37 crc kubenswrapper[4744]: I1008 09:24:37.911669 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t7n8" event={"ID":"417b66f9-47a8-4fc7-9d8e-e56c20bc5657","Type":"ContainerStarted","Data":"acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619"} Oct 08 09:24:37 crc kubenswrapper[4744]: I1008 09:24:37.932071 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9t7n8" podStartSLOduration=3.140492878 podStartE2EDuration="5.932047739s" podCreationTimestamp="2025-10-08 09:24:32 +0000 UTC" firstStartedPulling="2025-10-08 09:24:34.847612085 +0000 UTC m=+770.095257324" lastFinishedPulling="2025-10-08 09:24:37.639166926 +0000 UTC m=+772.886812185" observedRunningTime="2025-10-08 09:24:37.927974112 +0000 UTC m=+773.175619371" watchObservedRunningTime="2025-10-08 09:24:37.932047739 +0000 UTC m=+773.179692988" Oct 08 09:24:43 crc kubenswrapper[4744]: I1008 09:24:43.315789 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:43 crc kubenswrapper[4744]: I1008 09:24:43.316160 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:43 crc kubenswrapper[4744]: I1008 09:24:43.384797 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:44 crc kubenswrapper[4744]: I1008 09:24:44.044660 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:44 crc kubenswrapper[4744]: I1008 09:24:44.944327 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:45 crc kubenswrapper[4744]: I1008 09:24:45.014136 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:45 crc kubenswrapper[4744]: I1008 09:24:45.954056 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9t7n8"] Oct 08 09:24:46 crc kubenswrapper[4744]: I1008 09:24:46.002865 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9t7n8" podUID="417b66f9-47a8-4fc7-9d8e-e56c20bc5657" containerName="registry-server" containerID="cri-o://acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619" gracePeriod=2 Oct 08 09:24:46 crc kubenswrapper[4744]: I1008 09:24:46.556845 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:46 crc kubenswrapper[4744]: I1008 09:24:46.734526 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk9nz\" (UniqueName: \"kubernetes.io/projected/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-kube-api-access-lk9nz\") pod \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\" (UID: \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\") " Oct 08 09:24:46 crc kubenswrapper[4744]: I1008 09:24:46.734841 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-utilities\") pod \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\" (UID: \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\") " Oct 08 09:24:46 crc kubenswrapper[4744]: I1008 09:24:46.734995 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-catalog-content\") pod \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\" (UID: \"417b66f9-47a8-4fc7-9d8e-e56c20bc5657\") " Oct 08 09:24:46 crc kubenswrapper[4744]: I1008 09:24:46.736107 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-utilities" (OuterVolumeSpecName: "utilities") pod "417b66f9-47a8-4fc7-9d8e-e56c20bc5657" (UID: "417b66f9-47a8-4fc7-9d8e-e56c20bc5657"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:24:46 crc kubenswrapper[4744]: I1008 09:24:46.742330 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-kube-api-access-lk9nz" (OuterVolumeSpecName: "kube-api-access-lk9nz") pod "417b66f9-47a8-4fc7-9d8e-e56c20bc5657" (UID: "417b66f9-47a8-4fc7-9d8e-e56c20bc5657"). InnerVolumeSpecName "kube-api-access-lk9nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:24:46 crc kubenswrapper[4744]: I1008 09:24:46.789101 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "417b66f9-47a8-4fc7-9d8e-e56c20bc5657" (UID: "417b66f9-47a8-4fc7-9d8e-e56c20bc5657"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:24:46 crc kubenswrapper[4744]: I1008 09:24:46.836817 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk9nz\" (UniqueName: \"kubernetes.io/projected/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-kube-api-access-lk9nz\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:46 crc kubenswrapper[4744]: I1008 09:24:46.836866 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:46 crc kubenswrapper[4744]: I1008 09:24:46.836880 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/417b66f9-47a8-4fc7-9d8e-e56c20bc5657-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.023205 4744 generic.go:334] "Generic (PLEG): container finished" podID="417b66f9-47a8-4fc7-9d8e-e56c20bc5657" containerID="acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619" exitCode=0 Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.023307 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t7n8" event={"ID":"417b66f9-47a8-4fc7-9d8e-e56c20bc5657","Type":"ContainerDied","Data":"acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619"} Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.023403 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9t7n8" event={"ID":"417b66f9-47a8-4fc7-9d8e-e56c20bc5657","Type":"ContainerDied","Data":"d4f4e6f660edebb65beaf3ac10f533f3f5e08903b972b6545a6a05092925ed60"} Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.023470 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9t7n8" Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.023572 4744 scope.go:117] "RemoveContainer" containerID="acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619" Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.057915 4744 scope.go:117] "RemoveContainer" containerID="a8f3d2b19bde5f9c4939613ebd4646b2bf2ec1bc0b5ddf056ee8a0525fd46820" Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.083572 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9t7n8"] Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.089955 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9t7n8"] Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.098760 4744 scope.go:117] "RemoveContainer" containerID="59cb652825c3d7e2175db99d6c779385f57bf4fffb86b5f0b3dcb115d4b13824" Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.118899 4744 scope.go:117] "RemoveContainer" containerID="acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619" Oct 08 09:24:47 crc kubenswrapper[4744]: E1008 09:24:47.119436 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619\": container with ID starting with acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619 not found: ID does not exist" containerID="acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619" Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.119472 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619"} err="failed to get container status \"acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619\": rpc error: code = NotFound desc = could not find container \"acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619\": container with ID starting with acdd0964f7885ff9bbcda56cc7230251bb30d86a84e8b62860808156b8f72619 not found: ID does not exist" Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.119498 4744 scope.go:117] "RemoveContainer" containerID="a8f3d2b19bde5f9c4939613ebd4646b2bf2ec1bc0b5ddf056ee8a0525fd46820" Oct 08 09:24:47 crc kubenswrapper[4744]: E1008 09:24:47.120140 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8f3d2b19bde5f9c4939613ebd4646b2bf2ec1bc0b5ddf056ee8a0525fd46820\": container with ID starting with a8f3d2b19bde5f9c4939613ebd4646b2bf2ec1bc0b5ddf056ee8a0525fd46820 not found: ID does not exist" containerID="a8f3d2b19bde5f9c4939613ebd4646b2bf2ec1bc0b5ddf056ee8a0525fd46820" Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.120249 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8f3d2b19bde5f9c4939613ebd4646b2bf2ec1bc0b5ddf056ee8a0525fd46820"} err="failed to get container status \"a8f3d2b19bde5f9c4939613ebd4646b2bf2ec1bc0b5ddf056ee8a0525fd46820\": rpc error: code = NotFound desc = could not find container \"a8f3d2b19bde5f9c4939613ebd4646b2bf2ec1bc0b5ddf056ee8a0525fd46820\": container with ID starting with a8f3d2b19bde5f9c4939613ebd4646b2bf2ec1bc0b5ddf056ee8a0525fd46820 not found: ID does not exist" Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.120455 4744 scope.go:117] "RemoveContainer" containerID="59cb652825c3d7e2175db99d6c779385f57bf4fffb86b5f0b3dcb115d4b13824" Oct 08 09:24:47 crc kubenswrapper[4744]: E1008 09:24:47.121458 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59cb652825c3d7e2175db99d6c779385f57bf4fffb86b5f0b3dcb115d4b13824\": container with ID starting with 59cb652825c3d7e2175db99d6c779385f57bf4fffb86b5f0b3dcb115d4b13824 not found: ID does not exist" containerID="59cb652825c3d7e2175db99d6c779385f57bf4fffb86b5f0b3dcb115d4b13824" Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.121546 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59cb652825c3d7e2175db99d6c779385f57bf4fffb86b5f0b3dcb115d4b13824"} err="failed to get container status \"59cb652825c3d7e2175db99d6c779385f57bf4fffb86b5f0b3dcb115d4b13824\": rpc error: code = NotFound desc = could not find container \"59cb652825c3d7e2175db99d6c779385f57bf4fffb86b5f0b3dcb115d4b13824\": container with ID starting with 59cb652825c3d7e2175db99d6c779385f57bf4fffb86b5f0b3dcb115d4b13824 not found: ID does not exist" Oct 08 09:24:47 crc kubenswrapper[4744]: I1008 09:24:47.462757 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="417b66f9-47a8-4fc7-9d8e-e56c20bc5657" path="/var/lib/kubelet/pods/417b66f9-47a8-4fc7-9d8e-e56c20bc5657/volumes" Oct 08 09:24:48 crc kubenswrapper[4744]: I1008 09:24:48.548622 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b7845"] Oct 08 09:24:48 crc kubenswrapper[4744]: I1008 09:24:48.548989 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b7845" podUID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" containerName="registry-server" containerID="cri-o://8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98" gracePeriod=2 Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.027550 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.038254 4744 generic.go:334] "Generic (PLEG): container finished" podID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" containerID="8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98" exitCode=0 Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.038301 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b7845" event={"ID":"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6","Type":"ContainerDied","Data":"8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98"} Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.038310 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b7845" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.038333 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b7845" event={"ID":"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6","Type":"ContainerDied","Data":"6668a5a8fbfa30e4ef5074d3a8faa8110cee2b2cb0f9206c51c9209a8b101132"} Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.038357 4744 scope.go:117] "RemoveContainer" containerID="8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.063531 4744 scope.go:117] "RemoveContainer" containerID="d4071b39b3d98698df0f2858ad67c741edafdc72290671dd207031151826ec7c" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.089687 4744 scope.go:117] "RemoveContainer" containerID="b312b4b496f4889472ec7877fe31f2410ff3ae322cc1b679cc053a5ed13ac344" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.107028 4744 scope.go:117] "RemoveContainer" containerID="8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98" Oct 08 09:24:49 crc kubenswrapper[4744]: E1008 09:24:49.107572 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98\": container with ID starting with 8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98 not found: ID does not exist" containerID="8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.107623 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98"} err="failed to get container status \"8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98\": rpc error: code = NotFound desc = could not find container \"8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98\": container with ID starting with 8f9b5cd9f6ddcd2ebe2a93563c697ffbf39b3acc3b1e622eb364d0208dab9b98 not found: ID does not exist" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.107663 4744 scope.go:117] "RemoveContainer" containerID="d4071b39b3d98698df0f2858ad67c741edafdc72290671dd207031151826ec7c" Oct 08 09:24:49 crc kubenswrapper[4744]: E1008 09:24:49.108392 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4071b39b3d98698df0f2858ad67c741edafdc72290671dd207031151826ec7c\": container with ID starting with d4071b39b3d98698df0f2858ad67c741edafdc72290671dd207031151826ec7c not found: ID does not exist" containerID="d4071b39b3d98698df0f2858ad67c741edafdc72290671dd207031151826ec7c" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.108432 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4071b39b3d98698df0f2858ad67c741edafdc72290671dd207031151826ec7c"} err="failed to get container status \"d4071b39b3d98698df0f2858ad67c741edafdc72290671dd207031151826ec7c\": rpc error: code = NotFound desc = could not find container \"d4071b39b3d98698df0f2858ad67c741edafdc72290671dd207031151826ec7c\": container with ID starting with d4071b39b3d98698df0f2858ad67c741edafdc72290671dd207031151826ec7c not found: ID does not exist" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.108457 4744 scope.go:117] "RemoveContainer" containerID="b312b4b496f4889472ec7877fe31f2410ff3ae322cc1b679cc053a5ed13ac344" Oct 08 09:24:49 crc kubenswrapper[4744]: E1008 09:24:49.108778 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b312b4b496f4889472ec7877fe31f2410ff3ae322cc1b679cc053a5ed13ac344\": container with ID starting with b312b4b496f4889472ec7877fe31f2410ff3ae322cc1b679cc053a5ed13ac344 not found: ID does not exist" containerID="b312b4b496f4889472ec7877fe31f2410ff3ae322cc1b679cc053a5ed13ac344" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.108849 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b312b4b496f4889472ec7877fe31f2410ff3ae322cc1b679cc053a5ed13ac344"} err="failed to get container status \"b312b4b496f4889472ec7877fe31f2410ff3ae322cc1b679cc053a5ed13ac344\": rpc error: code = NotFound desc = could not find container \"b312b4b496f4889472ec7877fe31f2410ff3ae322cc1b679cc053a5ed13ac344\": container with ID starting with b312b4b496f4889472ec7877fe31f2410ff3ae322cc1b679cc053a5ed13ac344 not found: ID does not exist" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.168656 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-utilities\") pod \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\" (UID: \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\") " Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.168754 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-catalog-content\") pod \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\" (UID: \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\") " Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.168781 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69j8l\" (UniqueName: \"kubernetes.io/projected/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-kube-api-access-69j8l\") pod \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\" (UID: \"9bd561d0-b3a4-42cf-8e1a-ff87abe359b6\") " Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.171015 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-utilities" (OuterVolumeSpecName: "utilities") pod "9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" (UID: "9bd561d0-b3a4-42cf-8e1a-ff87abe359b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.173456 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-kube-api-access-69j8l" (OuterVolumeSpecName: "kube-api-access-69j8l") pod "9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" (UID: "9bd561d0-b3a4-42cf-8e1a-ff87abe359b6"). InnerVolumeSpecName "kube-api-access-69j8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.250352 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" (UID: "9bd561d0-b3a4-42cf-8e1a-ff87abe359b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.269865 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.269900 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.269920 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69j8l\" (UniqueName: \"kubernetes.io/projected/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6-kube-api-access-69j8l\") on node \"crc\" DevicePath \"\"" Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.367630 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b7845"] Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.371028 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b7845"] Oct 08 09:24:49 crc kubenswrapper[4744]: I1008 09:24:49.461201 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" path="/var/lib/kubelet/pods/9bd561d0-b3a4-42cf-8e1a-ff87abe359b6/volumes" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.761119 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-226vj"] Oct 08 09:25:03 crc kubenswrapper[4744]: E1008 09:25:03.761935 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="417b66f9-47a8-4fc7-9d8e-e56c20bc5657" containerName="extract-content" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.761951 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="417b66f9-47a8-4fc7-9d8e-e56c20bc5657" containerName="extract-content" Oct 08 09:25:03 crc kubenswrapper[4744]: E1008 09:25:03.761967 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" containerName="registry-server" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.761975 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" containerName="registry-server" Oct 08 09:25:03 crc kubenswrapper[4744]: E1008 09:25:03.761988 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="417b66f9-47a8-4fc7-9d8e-e56c20bc5657" containerName="registry-server" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.761997 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="417b66f9-47a8-4fc7-9d8e-e56c20bc5657" containerName="registry-server" Oct 08 09:25:03 crc kubenswrapper[4744]: E1008 09:25:03.762007 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="417b66f9-47a8-4fc7-9d8e-e56c20bc5657" containerName="extract-utilities" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.762014 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="417b66f9-47a8-4fc7-9d8e-e56c20bc5657" containerName="extract-utilities" Oct 08 09:25:03 crc kubenswrapper[4744]: E1008 09:25:03.762034 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" containerName="extract-content" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.762043 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" containerName="extract-content" Oct 08 09:25:03 crc kubenswrapper[4744]: E1008 09:25:03.762058 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" containerName="extract-utilities" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.762065 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" containerName="extract-utilities" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.762191 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="417b66f9-47a8-4fc7-9d8e-e56c20bc5657" containerName="registry-server" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.762213 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bd561d0-b3a4-42cf-8e1a-ff87abe359b6" containerName="registry-server" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.763225 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.784521 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-226vj"] Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.829114 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dae8c09-ed33-493e-baad-730bd1cdc511-utilities\") pod \"redhat-marketplace-226vj\" (UID: \"8dae8c09-ed33-493e-baad-730bd1cdc511\") " pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.829177 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dae8c09-ed33-493e-baad-730bd1cdc511-catalog-content\") pod \"redhat-marketplace-226vj\" (UID: \"8dae8c09-ed33-493e-baad-730bd1cdc511\") " pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.829217 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl59k\" (UniqueName: \"kubernetes.io/projected/8dae8c09-ed33-493e-baad-730bd1cdc511-kube-api-access-fl59k\") pod \"redhat-marketplace-226vj\" (UID: \"8dae8c09-ed33-493e-baad-730bd1cdc511\") " pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.930761 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dae8c09-ed33-493e-baad-730bd1cdc511-utilities\") pod \"redhat-marketplace-226vj\" (UID: \"8dae8c09-ed33-493e-baad-730bd1cdc511\") " pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.930831 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dae8c09-ed33-493e-baad-730bd1cdc511-catalog-content\") pod \"redhat-marketplace-226vj\" (UID: \"8dae8c09-ed33-493e-baad-730bd1cdc511\") " pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.930863 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl59k\" (UniqueName: \"kubernetes.io/projected/8dae8c09-ed33-493e-baad-730bd1cdc511-kube-api-access-fl59k\") pod \"redhat-marketplace-226vj\" (UID: \"8dae8c09-ed33-493e-baad-730bd1cdc511\") " pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.931444 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dae8c09-ed33-493e-baad-730bd1cdc511-catalog-content\") pod \"redhat-marketplace-226vj\" (UID: \"8dae8c09-ed33-493e-baad-730bd1cdc511\") " pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.931527 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dae8c09-ed33-493e-baad-730bd1cdc511-utilities\") pod \"redhat-marketplace-226vj\" (UID: \"8dae8c09-ed33-493e-baad-730bd1cdc511\") " pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:03 crc kubenswrapper[4744]: I1008 09:25:03.956971 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl59k\" (UniqueName: \"kubernetes.io/projected/8dae8c09-ed33-493e-baad-730bd1cdc511-kube-api-access-fl59k\") pod \"redhat-marketplace-226vj\" (UID: \"8dae8c09-ed33-493e-baad-730bd1cdc511\") " pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:04 crc kubenswrapper[4744]: I1008 09:25:04.081383 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:04 crc kubenswrapper[4744]: I1008 09:25:04.664514 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-226vj"] Oct 08 09:25:05 crc kubenswrapper[4744]: I1008 09:25:05.143880 4744 generic.go:334] "Generic (PLEG): container finished" podID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerID="27804c337514233ed1dd659773cff245ba0ec57691ca78b3851a4c9807d6470e" exitCode=0 Oct 08 09:25:05 crc kubenswrapper[4744]: I1008 09:25:05.143928 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-226vj" event={"ID":"8dae8c09-ed33-493e-baad-730bd1cdc511","Type":"ContainerDied","Data":"27804c337514233ed1dd659773cff245ba0ec57691ca78b3851a4c9807d6470e"} Oct 08 09:25:05 crc kubenswrapper[4744]: I1008 09:25:05.143958 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-226vj" event={"ID":"8dae8c09-ed33-493e-baad-730bd1cdc511","Type":"ContainerStarted","Data":"982c891c6c804b1d8917002d365332ce1d23185b8bbf171b6401d69f40fde797"} Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.151054 4744 generic.go:334] "Generic (PLEG): container finished" podID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerID="0cadae97da6e2218c0d41f376fc75918977e9300d9f3df63dfe57956d27b6c3a" exitCode=0 Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.151097 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-226vj" event={"ID":"8dae8c09-ed33-493e-baad-730bd1cdc511","Type":"ContainerDied","Data":"0cadae97da6e2218c0d41f376fc75918977e9300d9f3df63dfe57956d27b6c3a"} Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.807035 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb"] Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.808644 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.810920 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2"] Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.811927 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.818320 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-cstkj" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.856606 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-24zrk" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.869409 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fsbb\" (UniqueName: \"kubernetes.io/projected/6d6f4234-2b91-4399-8aa2-f1064fde501c-kube-api-access-2fsbb\") pod \"barbican-operator-controller-manager-658bdf4b74-nzjx2\" (UID: \"6d6f4234-2b91-4399-8aa2-f1064fde501c\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.869471 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n757w\" (UniqueName: \"kubernetes.io/projected/8a43b240-9fe3-4c10-afa9-7e78777e76d5-kube-api-access-n757w\") pod \"cinder-operator-controller-manager-7b7fb68549-csxmb\" (UID: \"8a43b240-9fe3-4c10-afa9-7e78777e76d5\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.888159 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2"] Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.896947 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl"] Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.898393 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.903507 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb"] Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.903783 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-g52q4" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.930828 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl"] Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.935844 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb"] Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.937335 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.949922 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-cq5rn" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.970910 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kp5z\" (UniqueName: \"kubernetes.io/projected/44cd0502-66fb-45ef-888a-da6273df7056-kube-api-access-2kp5z\") pod \"glance-operator-controller-manager-84b9b84486-f9zmb\" (UID: \"44cd0502-66fb-45ef-888a-da6273df7056\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.970989 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fsbb\" (UniqueName: \"kubernetes.io/projected/6d6f4234-2b91-4399-8aa2-f1064fde501c-kube-api-access-2fsbb\") pod \"barbican-operator-controller-manager-658bdf4b74-nzjx2\" (UID: \"6d6f4234-2b91-4399-8aa2-f1064fde501c\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.971031 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n757w\" (UniqueName: \"kubernetes.io/projected/8a43b240-9fe3-4c10-afa9-7e78777e76d5-kube-api-access-n757w\") pod \"cinder-operator-controller-manager-7b7fb68549-csxmb\" (UID: \"8a43b240-9fe3-4c10-afa9-7e78777e76d5\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.971055 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr848\" (UniqueName: \"kubernetes.io/projected/254b5ec4-bc16-4e8a-a61b-8a4a588dd629-kube-api-access-wr848\") pod \"designate-operator-controller-manager-85d5d9dd78-b9ntl\" (UID: \"254b5ec4-bc16-4e8a-a61b-8a4a588dd629\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl" Oct 08 09:25:06 crc kubenswrapper[4744]: I1008 09:25:06.993541 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.007127 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.015018 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.034490 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n757w\" (UniqueName: \"kubernetes.io/projected/8a43b240-9fe3-4c10-afa9-7e78777e76d5-kube-api-access-n757w\") pod \"cinder-operator-controller-manager-7b7fb68549-csxmb\" (UID: \"8a43b240-9fe3-4c10-afa9-7e78777e76d5\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.042877 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.042998 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-pv6z2" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.048567 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fsbb\" (UniqueName: \"kubernetes.io/projected/6d6f4234-2b91-4399-8aa2-f1064fde501c-kube-api-access-2fsbb\") pod \"barbican-operator-controller-manager-658bdf4b74-nzjx2\" (UID: \"6d6f4234-2b91-4399-8aa2-f1064fde501c\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.073552 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kp5z\" (UniqueName: \"kubernetes.io/projected/44cd0502-66fb-45ef-888a-da6273df7056-kube-api-access-2kp5z\") pod \"glance-operator-controller-manager-84b9b84486-f9zmb\" (UID: \"44cd0502-66fb-45ef-888a-da6273df7056\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.073632 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr848\" (UniqueName: \"kubernetes.io/projected/254b5ec4-bc16-4e8a-a61b-8a4a588dd629-kube-api-access-wr848\") pod \"designate-operator-controller-manager-85d5d9dd78-b9ntl\" (UID: \"254b5ec4-bc16-4e8a-a61b-8a4a588dd629\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.073709 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn7mf\" (UniqueName: \"kubernetes.io/projected/622e796c-d598-4721-89bd-c0cb3f83a1ad-kube-api-access-hn7mf\") pod \"heat-operator-controller-manager-858f76bbdd-m9prk\" (UID: \"622e796c-d598-4721-89bd-c0cb3f83a1ad\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.081739 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.082965 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.104379 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-5b9cd" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.106617 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.132651 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.138401 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.168216 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kp5z\" (UniqueName: \"kubernetes.io/projected/44cd0502-66fb-45ef-888a-da6273df7056-kube-api-access-2kp5z\") pod \"glance-operator-controller-manager-84b9b84486-f9zmb\" (UID: \"44cd0502-66fb-45ef-888a-da6273df7056\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.174766 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqbqs\" (UniqueName: \"kubernetes.io/projected/18a79d6e-9afe-43b5-8cf1-b5a991d2fd49-kube-api-access-pqbqs\") pod \"horizon-operator-controller-manager-7ffbcb7588-hsmrl\" (UID: \"18a79d6e-9afe-43b5-8cf1-b5a991d2fd49\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.174855 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn7mf\" (UniqueName: \"kubernetes.io/projected/622e796c-d598-4721-89bd-c0cb3f83a1ad-kube-api-access-hn7mf\") pod \"heat-operator-controller-manager-858f76bbdd-m9prk\" (UID: \"622e796c-d598-4721-89bd-c0cb3f83a1ad\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.176022 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr848\" (UniqueName: \"kubernetes.io/projected/254b5ec4-bc16-4e8a-a61b-8a4a588dd629-kube-api-access-wr848\") pod \"designate-operator-controller-manager-85d5d9dd78-b9ntl\" (UID: \"254b5ec4-bc16-4e8a-a61b-8a4a588dd629\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.186304 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.187264 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.195533 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.204753 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-226vj" event={"ID":"8dae8c09-ed33-493e-baad-730bd1cdc511","Type":"ContainerStarted","Data":"c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a"} Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.217712 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-d5k28" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.217801 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.246319 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.259924 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn7mf\" (UniqueName: \"kubernetes.io/projected/622e796c-d598-4721-89bd-c0cb3f83a1ad-kube-api-access-hn7mf\") pod \"heat-operator-controller-manager-858f76bbdd-m9prk\" (UID: \"622e796c-d598-4721-89bd-c0cb3f83a1ad\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.260708 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.278485 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqbqs\" (UniqueName: \"kubernetes.io/projected/18a79d6e-9afe-43b5-8cf1-b5a991d2fd49-kube-api-access-pqbqs\") pod \"horizon-operator-controller-manager-7ffbcb7588-hsmrl\" (UID: \"18a79d6e-9afe-43b5-8cf1-b5a991d2fd49\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.278592 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfb8d\" (UniqueName: \"kubernetes.io/projected/b2bb8369-17dd-4f3b-8e8d-4af7895d892c-kube-api-access-qfb8d\") pod \"infra-operator-controller-manager-656bcbd775-tbx59\" (UID: \"b2bb8369-17dd-4f3b-8e8d-4af7895d892c\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.278644 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b2bb8369-17dd-4f3b-8e8d-4af7895d892c-cert\") pod \"infra-operator-controller-manager-656bcbd775-tbx59\" (UID: \"b2bb8369-17dd-4f3b-8e8d-4af7895d892c\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.336476 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqbqs\" (UniqueName: \"kubernetes.io/projected/18a79d6e-9afe-43b5-8cf1-b5a991d2fd49-kube-api-access-pqbqs\") pod \"horizon-operator-controller-manager-7ffbcb7588-hsmrl\" (UID: \"18a79d6e-9afe-43b5-8cf1-b5a991d2fd49\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.351362 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.353260 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.358262 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-jkfbj" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.375502 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-226vj" podStartSLOduration=2.971877417 podStartE2EDuration="4.375481372s" podCreationTimestamp="2025-10-08 09:25:03 +0000 UTC" firstStartedPulling="2025-10-08 09:25:05.145591535 +0000 UTC m=+800.393236764" lastFinishedPulling="2025-10-08 09:25:06.54919549 +0000 UTC m=+801.796840719" observedRunningTime="2025-10-08 09:25:07.348933994 +0000 UTC m=+802.596579243" watchObservedRunningTime="2025-10-08 09:25:07.375481372 +0000 UTC m=+802.623126611" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.377688 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.378936 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.380941 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b2bb8369-17dd-4f3b-8e8d-4af7895d892c-cert\") pod \"infra-operator-controller-manager-656bcbd775-tbx59\" (UID: \"b2bb8369-17dd-4f3b-8e8d-4af7895d892c\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.381008 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgr6c\" (UniqueName: \"kubernetes.io/projected/6126b107-8c51-4467-9c40-e2c4b268cec5-kube-api-access-zgr6c\") pod \"ironic-operator-controller-manager-9c5c78d49-q9hcb\" (UID: \"6126b107-8c51-4467-9c40-e2c4b268cec5\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.381095 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfb8d\" (UniqueName: \"kubernetes.io/projected/b2bb8369-17dd-4f3b-8e8d-4af7895d892c-kube-api-access-qfb8d\") pod \"infra-operator-controller-manager-656bcbd775-tbx59\" (UID: \"b2bb8369-17dd-4f3b-8e8d-4af7895d892c\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" Oct 08 09:25:07 crc kubenswrapper[4744]: E1008 09:25:07.381511 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 08 09:25:07 crc kubenswrapper[4744]: E1008 09:25:07.381561 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2bb8369-17dd-4f3b-8e8d-4af7895d892c-cert podName:b2bb8369-17dd-4f3b-8e8d-4af7895d892c nodeName:}" failed. No retries permitted until 2025-10-08 09:25:07.881547105 +0000 UTC m=+803.129192344 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b2bb8369-17dd-4f3b-8e8d-4af7895d892c-cert") pod "infra-operator-controller-manager-656bcbd775-tbx59" (UID: "b2bb8369-17dd-4f3b-8e8d-4af7895d892c") : secret "infra-operator-webhook-server-cert" not found Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.396327 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.440592 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-4lrpx" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.442734 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.445244 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.450530 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.465499 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-88gjb" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.516778 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgr6c\" (UniqueName: \"kubernetes.io/projected/6126b107-8c51-4467-9c40-e2c4b268cec5-kube-api-access-zgr6c\") pod \"ironic-operator-controller-manager-9c5c78d49-q9hcb\" (UID: \"6126b107-8c51-4467-9c40-e2c4b268cec5\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.517120 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkj6s\" (UniqueName: \"kubernetes.io/projected/58decb1a-987e-4366-9bf8-ad4bf73e5969-kube-api-access-fkj6s\") pod \"manila-operator-controller-manager-5f67fbc655-lj9pr\" (UID: \"58decb1a-987e-4366-9bf8-ad4bf73e5969\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.517161 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56t76\" (UniqueName: \"kubernetes.io/projected/72a5635b-6920-4c2b-b91c-7621bc933959-kube-api-access-56t76\") pod \"keystone-operator-controller-manager-55b6b7c7b8-xjxx8\" (UID: \"72a5635b-6920-4c2b-b91c-7621bc933959\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.530466 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfb8d\" (UniqueName: \"kubernetes.io/projected/b2bb8369-17dd-4f3b-8e8d-4af7895d892c-kube-api-access-qfb8d\") pod \"infra-operator-controller-manager-656bcbd775-tbx59\" (UID: \"b2bb8369-17dd-4f3b-8e8d-4af7895d892c\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.625180 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkj6s\" (UniqueName: \"kubernetes.io/projected/58decb1a-987e-4366-9bf8-ad4bf73e5969-kube-api-access-fkj6s\") pod \"manila-operator-controller-manager-5f67fbc655-lj9pr\" (UID: \"58decb1a-987e-4366-9bf8-ad4bf73e5969\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.625326 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56t76\" (UniqueName: \"kubernetes.io/projected/72a5635b-6920-4c2b-b91c-7621bc933959-kube-api-access-56t76\") pod \"keystone-operator-controller-manager-55b6b7c7b8-xjxx8\" (UID: \"72a5635b-6920-4c2b-b91c-7621bc933959\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.625594 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.627356 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgr6c\" (UniqueName: \"kubernetes.io/projected/6126b107-8c51-4467-9c40-e2c4b268cec5-kube-api-access-zgr6c\") pod \"ironic-operator-controller-manager-9c5c78d49-q9hcb\" (UID: \"6126b107-8c51-4467-9c40-e2c4b268cec5\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.670503 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.685652 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkj6s\" (UniqueName: \"kubernetes.io/projected/58decb1a-987e-4366-9bf8-ad4bf73e5969-kube-api-access-fkj6s\") pod \"manila-operator-controller-manager-5f67fbc655-lj9pr\" (UID: \"58decb1a-987e-4366-9bf8-ad4bf73e5969\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.690571 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.714805 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.721659 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.730289 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56t76\" (UniqueName: \"kubernetes.io/projected/72a5635b-6920-4c2b-b91c-7621bc933959-kube-api-access-56t76\") pod \"keystone-operator-controller-manager-55b6b7c7b8-xjxx8\" (UID: \"72a5635b-6920-4c2b-b91c-7621bc933959\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.735228 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-4wtwf" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.760325 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.770457 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.788915 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.790349 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.798696 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.809078 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.852175 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-ltdvr" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.852719 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-kt6wg" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.855235 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.855329 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.860967 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.862047 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp6g4\" (UniqueName: \"kubernetes.io/projected/48158343-3842-44df-86c8-3a2f0e06a09c-kube-api-access-kp6g4\") pod \"neutron-operator-controller-manager-79d585cb66-xl4hw\" (UID: \"48158343-3842-44df-86c8-3a2f0e06a09c\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.862108 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m8tw\" (UniqueName: \"kubernetes.io/projected/2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34-kube-api-access-2m8tw\") pod \"nova-operator-controller-manager-5df598886f-5vz2c\" (UID: \"2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.862141 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdwls\" (UniqueName: \"kubernetes.io/projected/5fae63cb-c599-447b-bafb-2ed6bad836d3-kube-api-access-qdwls\") pod \"mariadb-operator-controller-manager-f9fb45f8f-4xvzd\" (UID: \"5fae63cb-c599-447b-bafb-2ed6bad836d3\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.894715 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.896589 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.918510 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-cch4w" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.920462 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.957329 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.965225 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b2bb8369-17dd-4f3b-8e8d-4af7895d892c-cert\") pod \"infra-operator-controller-manager-656bcbd775-tbx59\" (UID: \"b2bb8369-17dd-4f3b-8e8d-4af7895d892c\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.965322 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bxrg\" (UniqueName: \"kubernetes.io/projected/3eeb379f-9005-4908-ac9e-6dfb06405fb2-kube-api-access-5bxrg\") pod \"octavia-operator-controller-manager-69fdcfc5f5-crqx2\" (UID: \"3eeb379f-9005-4908-ac9e-6dfb06405fb2\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.965352 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp6g4\" (UniqueName: \"kubernetes.io/projected/48158343-3842-44df-86c8-3a2f0e06a09c-kube-api-access-kp6g4\") pod \"neutron-operator-controller-manager-79d585cb66-xl4hw\" (UID: \"48158343-3842-44df-86c8-3a2f0e06a09c\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.965425 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m8tw\" (UniqueName: \"kubernetes.io/projected/2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34-kube-api-access-2m8tw\") pod \"nova-operator-controller-manager-5df598886f-5vz2c\" (UID: \"2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.965463 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdwls\" (UniqueName: \"kubernetes.io/projected/5fae63cb-c599-447b-bafb-2ed6bad836d3-kube-api-access-qdwls\") pod \"mariadb-operator-controller-manager-f9fb45f8f-4xvzd\" (UID: \"5fae63cb-c599-447b-bafb-2ed6bad836d3\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.974955 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc"] Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.976468 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.990939 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b2bb8369-17dd-4f3b-8e8d-4af7895d892c-cert\") pod \"infra-operator-controller-manager-656bcbd775-tbx59\" (UID: \"b2bb8369-17dd-4f3b-8e8d-4af7895d892c\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" Oct 08 09:25:07 crc kubenswrapper[4744]: I1008 09:25:07.999704 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-x2hm8" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.000341 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.008712 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.020231 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.059468 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp6g4\" (UniqueName: \"kubernetes.io/projected/48158343-3842-44df-86c8-3a2f0e06a09c-kube-api-access-kp6g4\") pod \"neutron-operator-controller-manager-79d585cb66-xl4hw\" (UID: \"48158343-3842-44df-86c8-3a2f0e06a09c\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.064723 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.069072 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdwls\" (UniqueName: \"kubernetes.io/projected/5fae63cb-c599-447b-bafb-2ed6bad836d3-kube-api-access-qdwls\") pod \"mariadb-operator-controller-manager-f9fb45f8f-4xvzd\" (UID: \"5fae63cb-c599-447b-bafb-2ed6bad836d3\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.071628 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m8tw\" (UniqueName: \"kubernetes.io/projected/2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34-kube-api-access-2m8tw\") pod \"nova-operator-controller-manager-5df598886f-5vz2c\" (UID: \"2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.074627 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5096473-ff21-474e-b7fc-03196e4ee5a8-cert\") pod \"openstack-baremetal-operator-controller-manager-747747dfccwc6fw\" (UID: \"c5096473-ff21-474e-b7fc-03196e4ee5a8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.074720 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkk6z\" (UniqueName: \"kubernetes.io/projected/c5096473-ff21-474e-b7fc-03196e4ee5a8-kube-api-access-xkk6z\") pod \"openstack-baremetal-operator-controller-manager-747747dfccwc6fw\" (UID: \"c5096473-ff21-474e-b7fc-03196e4ee5a8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.074749 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bxrg\" (UniqueName: \"kubernetes.io/projected/3eeb379f-9005-4908-ac9e-6dfb06405fb2-kube-api-access-5bxrg\") pod \"octavia-operator-controller-manager-69fdcfc5f5-crqx2\" (UID: \"3eeb379f-9005-4908-ac9e-6dfb06405fb2\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.074801 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn48k\" (UniqueName: \"kubernetes.io/projected/e991d608-9aa2-4028-8080-a6bd7dca66c8-kube-api-access-kn48k\") pod \"ovn-operator-controller-manager-79db49b9fb-qppfc\" (UID: \"e991d608-9aa2-4028-8080-a6bd7dca66c8\") " pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.083718 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.100064 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.101854 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.134486 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.135997 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.156570 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-88q4q" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.162647 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-wd6fl" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.175267 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-shb79" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.175564 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkk6z\" (UniqueName: \"kubernetes.io/projected/c5096473-ff21-474e-b7fc-03196e4ee5a8-kube-api-access-xkk6z\") pod \"openstack-baremetal-operator-controller-manager-747747dfccwc6fw\" (UID: \"c5096473-ff21-474e-b7fc-03196e4ee5a8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.175643 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hqvl\" (UniqueName: \"kubernetes.io/projected/663cead5-7b5e-4940-94f9-5974f0424eda-kube-api-access-7hqvl\") pod \"swift-operator-controller-manager-db6d7f97b-frfr7\" (UID: \"663cead5-7b5e-4940-94f9-5974f0424eda\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.175665 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn48k\" (UniqueName: \"kubernetes.io/projected/e991d608-9aa2-4028-8080-a6bd7dca66c8-kube-api-access-kn48k\") pod \"ovn-operator-controller-manager-79db49b9fb-qppfc\" (UID: \"e991d608-9aa2-4028-8080-a6bd7dca66c8\") " pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.175699 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5096473-ff21-474e-b7fc-03196e4ee5a8-cert\") pod \"openstack-baremetal-operator-controller-manager-747747dfccwc6fw\" (UID: \"c5096473-ff21-474e-b7fc-03196e4ee5a8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" Oct 08 09:25:08 crc kubenswrapper[4744]: E1008 09:25:08.175856 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 09:25:08 crc kubenswrapper[4744]: E1008 09:25:08.175901 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c5096473-ff21-474e-b7fc-03196e4ee5a8-cert podName:c5096473-ff21-474e-b7fc-03196e4ee5a8 nodeName:}" failed. No retries permitted until 2025-10-08 09:25:08.675882745 +0000 UTC m=+803.923527984 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c5096473-ff21-474e-b7fc-03196e4ee5a8-cert") pod "openstack-baremetal-operator-controller-manager-747747dfccwc6fw" (UID: "c5096473-ff21-474e-b7fc-03196e4ee5a8") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.178292 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.181216 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.181749 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.205450 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.207103 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.247317 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.248645 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.251470 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.277182 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9txs\" (UniqueName: \"kubernetes.io/projected/3a897464-5474-43db-99d5-35f691f64220-kube-api-access-g9txs\") pod \"placement-operator-controller-manager-68b6c87b68-lbvtv\" (UID: \"3a897464-5474-43db-99d5-35f691f64220\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.277264 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hqvl\" (UniqueName: \"kubernetes.io/projected/663cead5-7b5e-4940-94f9-5974f0424eda-kube-api-access-7hqvl\") pod \"swift-operator-controller-manager-db6d7f97b-frfr7\" (UID: \"663cead5-7b5e-4940-94f9-5974f0424eda\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.277348 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwkpp\" (UniqueName: \"kubernetes.io/projected/f3456365-6f0d-4087-b753-40292c80bf12-kube-api-access-rwkpp\") pod \"telemetry-operator-controller-manager-76796d4c6b-gv4tw\" (UID: \"f3456365-6f0d-4087-b753-40292c80bf12\") " pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.311986 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-hqvm9" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.312210 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bxrg\" (UniqueName: \"kubernetes.io/projected/3eeb379f-9005-4908-ac9e-6dfb06405fb2-kube-api-access-5bxrg\") pod \"octavia-operator-controller-manager-69fdcfc5f5-crqx2\" (UID: \"3eeb379f-9005-4908-ac9e-6dfb06405fb2\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.312656 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn48k\" (UniqueName: \"kubernetes.io/projected/e991d608-9aa2-4028-8080-a6bd7dca66c8-kube-api-access-kn48k\") pod \"ovn-operator-controller-manager-79db49b9fb-qppfc\" (UID: \"e991d608-9aa2-4028-8080-a6bd7dca66c8\") " pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.328771 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hqvl\" (UniqueName: \"kubernetes.io/projected/663cead5-7b5e-4940-94f9-5974f0424eda-kube-api-access-7hqvl\") pod \"swift-operator-controller-manager-db6d7f97b-frfr7\" (UID: \"663cead5-7b5e-4940-94f9-5974f0424eda\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.361004 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkk6z\" (UniqueName: \"kubernetes.io/projected/c5096473-ff21-474e-b7fc-03196e4ee5a8-kube-api-access-xkk6z\") pod \"openstack-baremetal-operator-controller-manager-747747dfccwc6fw\" (UID: \"c5096473-ff21-474e-b7fc-03196e4ee5a8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.370871 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.388333 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwkpp\" (UniqueName: \"kubernetes.io/projected/f3456365-6f0d-4087-b753-40292c80bf12-kube-api-access-rwkpp\") pod \"telemetry-operator-controller-manager-76796d4c6b-gv4tw\" (UID: \"f3456365-6f0d-4087-b753-40292c80bf12\") " pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.388432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9txs\" (UniqueName: \"kubernetes.io/projected/3a897464-5474-43db-99d5-35f691f64220-kube-api-access-g9txs\") pod \"placement-operator-controller-manager-68b6c87b68-lbvtv\" (UID: \"3a897464-5474-43db-99d5-35f691f64220\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.422785 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.460294 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9txs\" (UniqueName: \"kubernetes.io/projected/3a897464-5474-43db-99d5-35f691f64220-kube-api-access-g9txs\") pod \"placement-operator-controller-manager-68b6c87b68-lbvtv\" (UID: \"3a897464-5474-43db-99d5-35f691f64220\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.465811 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.489814 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwkpp\" (UniqueName: \"kubernetes.io/projected/f3456365-6f0d-4087-b753-40292c80bf12-kube-api-access-rwkpp\") pod \"telemetry-operator-controller-manager-76796d4c6b-gv4tw\" (UID: \"f3456365-6f0d-4087-b753-40292c80bf12\") " pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.509231 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.510574 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.525045 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-5h8wn" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.596584 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-56c698c775-bf9cm"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.625953 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.632049 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.640393 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.648565 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.649856 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-2g8cf" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.680162 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.742077 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5096473-ff21-474e-b7fc-03196e4ee5a8-cert\") pod \"openstack-baremetal-operator-controller-manager-747747dfccwc6fw\" (UID: \"c5096473-ff21-474e-b7fc-03196e4ee5a8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.742854 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msxpj\" (UniqueName: \"kubernetes.io/projected/60f0c455-812d-40fe-b1b6-5aa75c34753c-kube-api-access-msxpj\") pod \"watcher-operator-controller-manager-7794bc6bd-q884x\" (UID: \"60f0c455-812d-40fe-b1b6-5aa75c34753c\") " pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.743402 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw" Oct 08 09:25:08 crc kubenswrapper[4744]: E1008 09:25:08.744064 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 09:25:08 crc kubenswrapper[4744]: E1008 09:25:08.744195 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c5096473-ff21-474e-b7fc-03196e4ee5a8-cert podName:c5096473-ff21-474e-b7fc-03196e4ee5a8 nodeName:}" failed. No retries permitted until 2025-10-08 09:25:09.74417653 +0000 UTC m=+804.991821769 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c5096473-ff21-474e-b7fc-03196e4ee5a8-cert") pod "openstack-baremetal-operator-controller-manager-747747dfccwc6fw" (UID: "c5096473-ff21-474e-b7fc-03196e4ee5a8") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.752108 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56c698c775-bf9cm"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.828605 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.862154 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msxpj\" (UniqueName: \"kubernetes.io/projected/60f0c455-812d-40fe-b1b6-5aa75c34753c-kube-api-access-msxpj\") pod \"watcher-operator-controller-manager-7794bc6bd-q884x\" (UID: \"60f0c455-812d-40fe-b1b6-5aa75c34753c\") " pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.894767 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msxpj\" (UniqueName: \"kubernetes.io/projected/60f0c455-812d-40fe-b1b6-5aa75c34753c-kube-api-access-msxpj\") pod \"watcher-operator-controller-manager-7794bc6bd-q884x\" (UID: \"60f0c455-812d-40fe-b1b6-5aa75c34753c\") " pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.929808 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz"] Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.931718 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.942821 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.943996 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-drkc7" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.965452 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bjp4\" (UniqueName: \"kubernetes.io/projected/12b5913f-74df-4ba6-8aa8-af5989327ddd-kube-api-access-6bjp4\") pod \"test-operator-controller-manager-56c698c775-bf9cm\" (UID: \"12b5913f-74df-4ba6-8aa8-af5989327ddd\") " pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" Oct 08 09:25:08 crc kubenswrapper[4744]: I1008 09:25:08.996991 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb"] Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.008057 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz"] Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.068177 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fdbcdf9-cab4-4d8a-9270-2a121c83dc04-cert\") pod \"openstack-operator-controller-manager-8bc6b8f5b-7q7gz\" (UID: \"6fdbcdf9-cab4-4d8a-9270-2a121c83dc04\") " pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.068449 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnqcd\" (UniqueName: \"kubernetes.io/projected/6fdbcdf9-cab4-4d8a-9270-2a121c83dc04-kube-api-access-qnqcd\") pod \"openstack-operator-controller-manager-8bc6b8f5b-7q7gz\" (UID: \"6fdbcdf9-cab4-4d8a-9270-2a121c83dc04\") " pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.068629 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bjp4\" (UniqueName: \"kubernetes.io/projected/12b5913f-74df-4ba6-8aa8-af5989327ddd-kube-api-access-6bjp4\") pod \"test-operator-controller-manager-56c698c775-bf9cm\" (UID: \"12b5913f-74df-4ba6-8aa8-af5989327ddd\") " pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.069318 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4"] Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.095900 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4"] Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.096194 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.104851 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk"] Oct 08 09:25:09 crc kubenswrapper[4744]: W1008 09:25:09.110782 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod622e796c_d598_4721_89bd_c0cb3f83a1ad.slice/crio-8bb7eba3d47076a585503cf15f32db56fb41cc332b74232c9cfe9d837c0c31ac WatchSource:0}: Error finding container 8bb7eba3d47076a585503cf15f32db56fb41cc332b74232c9cfe9d837c0c31ac: Status 404 returned error can't find the container with id 8bb7eba3d47076a585503cf15f32db56fb41cc332b74232c9cfe9d837c0c31ac Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.112190 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl"] Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.117293 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-sw5qm" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.169484 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bjp4\" (UniqueName: \"kubernetes.io/projected/12b5913f-74df-4ba6-8aa8-af5989327ddd-kube-api-access-6bjp4\") pod \"test-operator-controller-manager-56c698c775-bf9cm\" (UID: \"12b5913f-74df-4ba6-8aa8-af5989327ddd\") " pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.170495 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fdbcdf9-cab4-4d8a-9270-2a121c83dc04-cert\") pod \"openstack-operator-controller-manager-8bc6b8f5b-7q7gz\" (UID: \"6fdbcdf9-cab4-4d8a-9270-2a121c83dc04\") " pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.170543 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnqcd\" (UniqueName: \"kubernetes.io/projected/6fdbcdf9-cab4-4d8a-9270-2a121c83dc04-kube-api-access-qnqcd\") pod \"openstack-operator-controller-manager-8bc6b8f5b-7q7gz\" (UID: \"6fdbcdf9-cab4-4d8a-9270-2a121c83dc04\") " pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" Oct 08 09:25:09 crc kubenswrapper[4744]: E1008 09:25:09.171137 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 08 09:25:09 crc kubenswrapper[4744]: E1008 09:25:09.171192 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fdbcdf9-cab4-4d8a-9270-2a121c83dc04-cert podName:6fdbcdf9-cab4-4d8a-9270-2a121c83dc04 nodeName:}" failed. No retries permitted until 2025-10-08 09:25:09.671173062 +0000 UTC m=+804.918818301 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6fdbcdf9-cab4-4d8a-9270-2a121c83dc04-cert") pod "openstack-operator-controller-manager-8bc6b8f5b-7q7gz" (UID: "6fdbcdf9-cab4-4d8a-9270-2a121c83dc04") : secret "webhook-server-cert" not found Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.186002 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x" Oct 08 09:25:09 crc kubenswrapper[4744]: W1008 09:25:09.197836 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18a79d6e_9afe_43b5_8cf1_b5a991d2fd49.slice/crio-f05f66b77c18911a2551259f4d97e29a4c6e4a03c5d96ce3dd7686ac1779af15 WatchSource:0}: Error finding container f05f66b77c18911a2551259f4d97e29a4c6e4a03c5d96ce3dd7686ac1779af15: Status 404 returned error can't find the container with id f05f66b77c18911a2551259f4d97e29a4c6e4a03c5d96ce3dd7686ac1779af15 Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.229748 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnqcd\" (UniqueName: \"kubernetes.io/projected/6fdbcdf9-cab4-4d8a-9270-2a121c83dc04-kube-api-access-qnqcd\") pod \"openstack-operator-controller-manager-8bc6b8f5b-7q7gz\" (UID: \"6fdbcdf9-cab4-4d8a-9270-2a121c83dc04\") " pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.273862 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5hfv\" (UniqueName: \"kubernetes.io/projected/5bf79b31-d3bc-4422-9302-ac7a55a1d95d-kube-api-access-v5hfv\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-ztld4\" (UID: \"5bf79b31-d3bc-4422-9302-ac7a55a1d95d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.300843 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8"] Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.327655 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb" event={"ID":"8a43b240-9fe3-4c10-afa9-7e78777e76d5","Type":"ContainerStarted","Data":"80f0560d3f1f4e5eac778ae527ac2d0f463f1077c4c2d11f9182d9f83c4f3e9e"} Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.333915 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.361319 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb" event={"ID":"44cd0502-66fb-45ef-888a-da6273df7056","Type":"ContainerStarted","Data":"c93aa24651c07e032a20088f3665872dbbcb22ca17efaa7c6ab42905669a6cd1"} Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.372654 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" event={"ID":"622e796c-d598-4721-89bd-c0cb3f83a1ad","Type":"ContainerStarted","Data":"8bb7eba3d47076a585503cf15f32db56fb41cc332b74232c9cfe9d837c0c31ac"} Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.379187 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5hfv\" (UniqueName: \"kubernetes.io/projected/5bf79b31-d3bc-4422-9302-ac7a55a1d95d-kube-api-access-v5hfv\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-ztld4\" (UID: \"5bf79b31-d3bc-4422-9302-ac7a55a1d95d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.392493 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" event={"ID":"18a79d6e-9afe-43b5-8cf1-b5a991d2fd49","Type":"ContainerStarted","Data":"f05f66b77c18911a2551259f4d97e29a4c6e4a03c5d96ce3dd7686ac1779af15"} Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.393507 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2"] Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.400707 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl" event={"ID":"254b5ec4-bc16-4e8a-a61b-8a4a588dd629","Type":"ContainerStarted","Data":"327cf04fcd583729eb86485f4c51d15a9ee5782ad62b382ecfa53733d110f911"} Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.409529 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5hfv\" (UniqueName: \"kubernetes.io/projected/5bf79b31-d3bc-4422-9302-ac7a55a1d95d-kube-api-access-v5hfv\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-ztld4\" (UID: \"5bf79b31-d3bc-4422-9302-ac7a55a1d95d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4" Oct 08 09:25:09 crc kubenswrapper[4744]: W1008 09:25:09.443897 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72a5635b_6920_4c2b_b91c_7621bc933959.slice/crio-c58dddbada39a2f29af3c319c95e4165c8cdf342bfe6f7d245fe02a609c46dcf WatchSource:0}: Error finding container c58dddbada39a2f29af3c319c95e4165c8cdf342bfe6f7d245fe02a609c46dcf: Status 404 returned error can't find the container with id c58dddbada39a2f29af3c319c95e4165c8cdf342bfe6f7d245fe02a609c46dcf Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.589285 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4" Oct 08 09:25:09 crc kubenswrapper[4744]: W1008 09:25:09.635358 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58decb1a_987e_4366_9bf8_ad4bf73e5969.slice/crio-c77add7e3a11112be3ec548180f7f6756218953c56c8c597d8a9f7810f5e0af1 WatchSource:0}: Error finding container c77add7e3a11112be3ec548180f7f6756218953c56c8c597d8a9f7810f5e0af1: Status 404 returned error can't find the container with id c77add7e3a11112be3ec548180f7f6756218953c56c8c597d8a9f7810f5e0af1 Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.640508 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr"] Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.671216 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb"] Oct 08 09:25:09 crc kubenswrapper[4744]: W1008 09:25:09.679573 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6126b107_8c51_4467_9c40_e2c4b268cec5.slice/crio-62eb933b0751a456cc4d3c6bf33d3f85f08604dc49901d6d6af29979bb08acc7 WatchSource:0}: Error finding container 62eb933b0751a456cc4d3c6bf33d3f85f08604dc49901d6d6af29979bb08acc7: Status 404 returned error can't find the container with id 62eb933b0751a456cc4d3c6bf33d3f85f08604dc49901d6d6af29979bb08acc7 Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.690052 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fdbcdf9-cab4-4d8a-9270-2a121c83dc04-cert\") pod \"openstack-operator-controller-manager-8bc6b8f5b-7q7gz\" (UID: \"6fdbcdf9-cab4-4d8a-9270-2a121c83dc04\") " pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" Oct 08 09:25:09 crc kubenswrapper[4744]: E1008 09:25:09.690303 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 08 09:25:09 crc kubenswrapper[4744]: E1008 09:25:09.690382 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6fdbcdf9-cab4-4d8a-9270-2a121c83dc04-cert podName:6fdbcdf9-cab4-4d8a-9270-2a121c83dc04 nodeName:}" failed. No retries permitted until 2025-10-08 09:25:10.690352785 +0000 UTC m=+805.937998024 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6fdbcdf9-cab4-4d8a-9270-2a121c83dc04-cert") pod "openstack-operator-controller-manager-8bc6b8f5b-7q7gz" (UID: "6fdbcdf9-cab4-4d8a-9270-2a121c83dc04") : secret "webhook-server-cert" not found Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.793220 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5096473-ff21-474e-b7fc-03196e4ee5a8-cert\") pod \"openstack-baremetal-operator-controller-manager-747747dfccwc6fw\" (UID: \"c5096473-ff21-474e-b7fc-03196e4ee5a8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.798022 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5096473-ff21-474e-b7fc-03196e4ee5a8-cert\") pod \"openstack-baremetal-operator-controller-manager-747747dfccwc6fw\" (UID: \"c5096473-ff21-474e-b7fc-03196e4ee5a8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.957440 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd"] Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.976801 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw"] Oct 08 09:25:09 crc kubenswrapper[4744]: I1008 09:25:09.983397 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c"] Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:09.996164 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.119152 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59"] Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.190468 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2"] Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.204777 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7"] Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.239048 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv"] Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.246312 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc"] Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.405243 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw"] Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.448443 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x"] Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.463877 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56c698c775-bf9cm"] Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.474559 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2" event={"ID":"3eeb379f-9005-4908-ac9e-6dfb06405fb2","Type":"ContainerStarted","Data":"e6d0b8fea5d15d2596c5384982b312f345c4575b93a1dd97fe93151768fa0ee1"} Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.483139 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" event={"ID":"b2bb8369-17dd-4f3b-8e8d-4af7895d892c","Type":"ContainerStarted","Data":"17576a1f287500335bbc0531a25063d7881d1b3f6af51d2e07cd35f0e71d0df9"} Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.490445 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc" event={"ID":"e991d608-9aa2-4028-8080-a6bd7dca66c8","Type":"ContainerStarted","Data":"185ace58f2407029f5c258f4b324167a34b46b12eff3e63b7252c2775c6fac21"} Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.493205 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv" event={"ID":"3a897464-5474-43db-99d5-35f691f64220","Type":"ContainerStarted","Data":"26044268fc7d12b60d68f83dedeafed8b349320bb0dcb09bb159d126de9171df"} Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.495406 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" event={"ID":"48158343-3842-44df-86c8-3a2f0e06a09c","Type":"ContainerStarted","Data":"eeadd229ce28747c3afa52c233abd96a1840ff9655326f8d7a7bfa12d8074147"} Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.503084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr" event={"ID":"58decb1a-987e-4366-9bf8-ad4bf73e5969","Type":"ContainerStarted","Data":"c77add7e3a11112be3ec548180f7f6756218953c56c8c597d8a9f7810f5e0af1"} Oct 08 09:25:10 crc kubenswrapper[4744]: E1008 09:25:10.504183 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:efa8fb78cffb573d299ffcc7bab1099affd2dbbab222152092b313074306e0a9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6bjp4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-56c698c775-bf9cm_openstack-operators(12b5913f-74df-4ba6-8aa8-af5989327ddd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.527068 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2" event={"ID":"6d6f4234-2b91-4399-8aa2-f1064fde501c","Type":"ContainerStarted","Data":"72f3cf682096101979138a6abe3eccfc32427debdc5f83e60ffd0c2f62c38050"} Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.528210 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c" event={"ID":"2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34","Type":"ContainerStarted","Data":"6be15ff3644ac25dbc2dcf34e0a3333d69b190e10bcbf849cd9bd8bae740173e"} Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.528278 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4"] Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.529747 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" event={"ID":"663cead5-7b5e-4940-94f9-5974f0424eda","Type":"ContainerStarted","Data":"0ef5c0730fd90d82e3de5e91c29a691af161b742e1ce9639a2055e391e17e780"} Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.539835 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" event={"ID":"5fae63cb-c599-447b-bafb-2ed6bad836d3","Type":"ContainerStarted","Data":"4e222fb097581620b7ddefc343ac99924f6d976a4d0543776f0bae395437b2bf"} Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.553172 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" event={"ID":"72a5635b-6920-4c2b-b91c-7621bc933959","Type":"ContainerStarted","Data":"c58dddbada39a2f29af3c319c95e4165c8cdf342bfe6f7d245fe02a609c46dcf"} Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.556891 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb" event={"ID":"6126b107-8c51-4467-9c40-e2c4b268cec5","Type":"ContainerStarted","Data":"62eb933b0751a456cc4d3c6bf33d3f85f08604dc49901d6d6af29979bb08acc7"} Oct 08 09:25:10 crc kubenswrapper[4744]: E1008 09:25:10.573157 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v5hfv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-ztld4_openstack-operators(5bf79b31-d3bc-4422-9302-ac7a55a1d95d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 09:25:10 crc kubenswrapper[4744]: E1008 09:25:10.580592 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4" podUID="5bf79b31-d3bc-4422-9302-ac7a55a1d95d" Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.749027 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fdbcdf9-cab4-4d8a-9270-2a121c83dc04-cert\") pod \"openstack-operator-controller-manager-8bc6b8f5b-7q7gz\" (UID: \"6fdbcdf9-cab4-4d8a-9270-2a121c83dc04\") " pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.757221 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6fdbcdf9-cab4-4d8a-9270-2a121c83dc04-cert\") pod \"openstack-operator-controller-manager-8bc6b8f5b-7q7gz\" (UID: \"6fdbcdf9-cab4-4d8a-9270-2a121c83dc04\") " pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" Oct 08 09:25:10 crc kubenswrapper[4744]: E1008 09:25:10.809939 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" podUID="12b5913f-74df-4ba6-8aa8-af5989327ddd" Oct 08 09:25:10 crc kubenswrapper[4744]: I1008 09:25:10.818823 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw"] Oct 08 09:25:11 crc kubenswrapper[4744]: I1008 09:25:11.020924 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" Oct 08 09:25:11 crc kubenswrapper[4744]: I1008 09:25:11.605276 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw" event={"ID":"f3456365-6f0d-4087-b753-40292c80bf12","Type":"ContainerStarted","Data":"db5d694bdd0cbbf06676980f7e8160af0235808e653b98f72c192d840f0d6528"} Oct 08 09:25:11 crc kubenswrapper[4744]: I1008 09:25:11.626875 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" event={"ID":"c5096473-ff21-474e-b7fc-03196e4ee5a8","Type":"ContainerStarted","Data":"6dbfa95226ff4576abc9eb177df2dedb397e0183d959de435d772a932c0d87d2"} Oct 08 09:25:11 crc kubenswrapper[4744]: I1008 09:25:11.648422 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x" event={"ID":"60f0c455-812d-40fe-b1b6-5aa75c34753c","Type":"ContainerStarted","Data":"58f368ff8ff87059a3b3df45d25dd52d6c48f8c507012d0045aaa4f8b27993e6"} Oct 08 09:25:11 crc kubenswrapper[4744]: I1008 09:25:11.654328 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4" event={"ID":"5bf79b31-d3bc-4422-9302-ac7a55a1d95d","Type":"ContainerStarted","Data":"be2ff1519d1c723f1765ad83d9ae89fbf5c0821337c404da6adfaf3c6667de3d"} Oct 08 09:25:11 crc kubenswrapper[4744]: E1008 09:25:11.660924 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4" podUID="5bf79b31-d3bc-4422-9302-ac7a55a1d95d" Oct 08 09:25:11 crc kubenswrapper[4744]: I1008 09:25:11.746618 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" event={"ID":"12b5913f-74df-4ba6-8aa8-af5989327ddd","Type":"ContainerStarted","Data":"f00f01aaeb21a80c9c02ac9b0939ecc295fc04baadd294525ba1c3805d5f1459"} Oct 08 09:25:11 crc kubenswrapper[4744]: I1008 09:25:11.746806 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" event={"ID":"12b5913f-74df-4ba6-8aa8-af5989327ddd","Type":"ContainerStarted","Data":"1fec3bd9bb287a6bbd30cb241e88a83a7c77a98a0a0001ec34e62bb26679ed71"} Oct 08 09:25:11 crc kubenswrapper[4744]: E1008 09:25:11.759916 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:efa8fb78cffb573d299ffcc7bab1099affd2dbbab222152092b313074306e0a9\\\"\"" pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" podUID="12b5913f-74df-4ba6-8aa8-af5989327ddd" Oct 08 09:25:11 crc kubenswrapper[4744]: I1008 09:25:11.919467 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz"] Oct 08 09:25:11 crc kubenswrapper[4744]: W1008 09:25:11.941492 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fdbcdf9_cab4_4d8a_9270_2a121c83dc04.slice/crio-b16b358645d22c7f8b3942280ab4c961caf1f3be24203447fa38ad88b5a26427 WatchSource:0}: Error finding container b16b358645d22c7f8b3942280ab4c961caf1f3be24203447fa38ad88b5a26427: Status 404 returned error can't find the container with id b16b358645d22c7f8b3942280ab4c961caf1f3be24203447fa38ad88b5a26427 Oct 08 09:25:12 crc kubenswrapper[4744]: I1008 09:25:12.797785 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" event={"ID":"6fdbcdf9-cab4-4d8a-9270-2a121c83dc04","Type":"ContainerStarted","Data":"8cd7791683dcb76aeb626b857f125ad64981896043a74cfab4f89f086b0da8e7"} Oct 08 09:25:12 crc kubenswrapper[4744]: I1008 09:25:12.798265 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" Oct 08 09:25:12 crc kubenswrapper[4744]: I1008 09:25:12.798277 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" event={"ID":"6fdbcdf9-cab4-4d8a-9270-2a121c83dc04","Type":"ContainerStarted","Data":"a78c1ab8cf37190df63c4d16f958f4b755b4bf53697e8455dc2a62f9f4443cfe"} Oct 08 09:25:12 crc kubenswrapper[4744]: I1008 09:25:12.798290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" event={"ID":"6fdbcdf9-cab4-4d8a-9270-2a121c83dc04","Type":"ContainerStarted","Data":"b16b358645d22c7f8b3942280ab4c961caf1f3be24203447fa38ad88b5a26427"} Oct 08 09:25:12 crc kubenswrapper[4744]: E1008 09:25:12.800072 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:efa8fb78cffb573d299ffcc7bab1099affd2dbbab222152092b313074306e0a9\\\"\"" pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" podUID="12b5913f-74df-4ba6-8aa8-af5989327ddd" Oct 08 09:25:12 crc kubenswrapper[4744]: E1008 09:25:12.800405 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4" podUID="5bf79b31-d3bc-4422-9302-ac7a55a1d95d" Oct 08 09:25:12 crc kubenswrapper[4744]: I1008 09:25:12.889960 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" podStartSLOduration=4.889919008 podStartE2EDuration="4.889919008s" podCreationTimestamp="2025-10-08 09:25:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:25:12.862894956 +0000 UTC m=+808.110540205" watchObservedRunningTime="2025-10-08 09:25:12.889919008 +0000 UTC m=+808.137564237" Oct 08 09:25:14 crc kubenswrapper[4744]: I1008 09:25:14.082303 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:14 crc kubenswrapper[4744]: I1008 09:25:14.082811 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:14 crc kubenswrapper[4744]: I1008 09:25:14.168928 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:14 crc kubenswrapper[4744]: I1008 09:25:14.893429 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:14 crc kubenswrapper[4744]: I1008 09:25:14.957053 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-226vj"] Oct 08 09:25:16 crc kubenswrapper[4744]: I1008 09:25:16.857401 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-226vj" podUID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerName="registry-server" containerID="cri-o://c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a" gracePeriod=2 Oct 08 09:25:17 crc kubenswrapper[4744]: I1008 09:25:17.866427 4744 generic.go:334] "Generic (PLEG): container finished" podID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerID="c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a" exitCode=0 Oct 08 09:25:17 crc kubenswrapper[4744]: I1008 09:25:17.866480 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-226vj" event={"ID":"8dae8c09-ed33-493e-baad-730bd1cdc511","Type":"ContainerDied","Data":"c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a"} Oct 08 09:25:21 crc kubenswrapper[4744]: I1008 09:25:21.028201 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-8bc6b8f5b-7q7gz" Oct 08 09:25:24 crc kubenswrapper[4744]: E1008 09:25:24.083005 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a is running failed: container process not found" containerID="c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 09:25:24 crc kubenswrapper[4744]: E1008 09:25:24.084574 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a is running failed: container process not found" containerID="c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 09:25:24 crc kubenswrapper[4744]: E1008 09:25:24.085850 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a is running failed: container process not found" containerID="c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 09:25:24 crc kubenswrapper[4744]: E1008 09:25:24.085918 4744 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-226vj" podUID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerName="registry-server" Oct 08 09:25:24 crc kubenswrapper[4744]: E1008 09:25:24.525828 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:582f7b1e411961b69f2e3c6b346aa25759b89f7720ed3fade1d363bf5d2dffc8" Oct 08 09:25:24 crc kubenswrapper[4744]: E1008 09:25:24.526063 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:582f7b1e411961b69f2e3c6b346aa25759b89f7720ed3fade1d363bf5d2dffc8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fkj6s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5f67fbc655-lj9pr_openstack-operators(58decb1a-987e-4366-9bf8-ad4bf73e5969): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:25:24 crc kubenswrapper[4744]: E1008 09:25:24.998180 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:9d26476523320d70d6d457b91663e8c233ed320d77032a7c57a89ce1aedd3931" Oct 08 09:25:24 crc kubenswrapper[4744]: E1008 09:25:24.998764 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:9d26476523320d70d6d457b91663e8c233ed320d77032a7c57a89ce1aedd3931,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rwkpp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76796d4c6b-gv4tw_openstack-operators(f3456365-6f0d-4087-b753-40292c80bf12): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.107788 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jczf2"] Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.109274 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.125583 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jczf2"] Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.240695 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s4m9\" (UniqueName: \"kubernetes.io/projected/23976bfb-8e75-4752-99ef-250d8b0d1e9d-kube-api-access-7s4m9\") pod \"certified-operators-jczf2\" (UID: \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\") " pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.240823 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23976bfb-8e75-4752-99ef-250d8b0d1e9d-utilities\") pod \"certified-operators-jczf2\" (UID: \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\") " pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.240899 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23976bfb-8e75-4752-99ef-250d8b0d1e9d-catalog-content\") pod \"certified-operators-jczf2\" (UID: \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\") " pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.342938 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s4m9\" (UniqueName: \"kubernetes.io/projected/23976bfb-8e75-4752-99ef-250d8b0d1e9d-kube-api-access-7s4m9\") pod \"certified-operators-jczf2\" (UID: \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\") " pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.343050 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23976bfb-8e75-4752-99ef-250d8b0d1e9d-utilities\") pod \"certified-operators-jczf2\" (UID: \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\") " pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.343099 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23976bfb-8e75-4752-99ef-250d8b0d1e9d-catalog-content\") pod \"certified-operators-jczf2\" (UID: \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\") " pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.346807 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23976bfb-8e75-4752-99ef-250d8b0d1e9d-utilities\") pod \"certified-operators-jczf2\" (UID: \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\") " pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.346925 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23976bfb-8e75-4752-99ef-250d8b0d1e9d-catalog-content\") pod \"certified-operators-jczf2\" (UID: \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\") " pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.371575 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s4m9\" (UniqueName: \"kubernetes.io/projected/23976bfb-8e75-4752-99ef-250d8b0d1e9d-kube-api-access-7s4m9\") pod \"certified-operators-jczf2\" (UID: \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\") " pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:25 crc kubenswrapper[4744]: I1008 09:25:25.430805 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:27 crc kubenswrapper[4744]: E1008 09:25:27.502970 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492" Oct 08 09:25:27 crc kubenswrapper[4744]: E1008 09:25:27.503565 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qfb8d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-656bcbd775-tbx59_openstack-operators(b2bb8369-17dd-4f3b-8e8d-4af7895d892c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:25:28 crc kubenswrapper[4744]: E1008 09:25:28.098794 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167" Oct 08 09:25:28 crc kubenswrapper[4744]: E1008 09:25:28.099302 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qdwls,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-f9fb45f8f-4xvzd_openstack-operators(5fae63cb-c599-447b-bafb-2ed6bad836d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:25:29 crc kubenswrapper[4744]: E1008 09:25:29.292595 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e" Oct 08 09:25:29 crc kubenswrapper[4744]: E1008 09:25:29.292850 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7hqvl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-db6d7f97b-frfr7_openstack-operators(663cead5-7b5e-4940-94f9-5974f0424eda): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:25:29 crc kubenswrapper[4744]: E1008 09:25:29.843256 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2" Oct 08 09:25:29 crc kubenswrapper[4744]: E1008 09:25:29.843503 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kp6g4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-79d585cb66-xl4hw_openstack-operators(48158343-3842-44df-86c8-3a2f0e06a09c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:25:30 crc kubenswrapper[4744]: E1008 09:25:30.502484 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:063a7e65b4ba98f0506f269ff7525b446eae06a5ed4a61c18ffa33a886500867" Oct 08 09:25:30 crc kubenswrapper[4744]: E1008 09:25:30.502835 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:063a7e65b4ba98f0506f269ff7525b446eae06a5ed4a61c18ffa33a886500867,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pqbqs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-7ffbcb7588-hsmrl_openstack-operators(18a79d6e-9afe-43b5-8cf1-b5a991d2fd49): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:25:31 crc kubenswrapper[4744]: E1008 09:25:31.164297 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997" Oct 08 09:25:31 crc kubenswrapper[4744]: E1008 09:25:31.164646 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-56t76,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-55b6b7c7b8-xjxx8_openstack-operators(72a5635b-6920-4c2b-b91c-7621bc933959): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:25:33 crc kubenswrapper[4744]: E1008 09:25:33.948695 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:ec11cb8711bd1af22db3c84aa854349ee46191add3db45aecfabb1d8410c04d0" Oct 08 09:25:33 crc kubenswrapper[4744]: E1008 09:25:33.949111 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:ec11cb8711bd1af22db3c84aa854349ee46191add3db45aecfabb1d8410c04d0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hn7mf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-858f76bbdd-m9prk_openstack-operators(622e796c-d598-4721-89bd-c0cb3f83a1ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:25:34 crc kubenswrapper[4744]: E1008 09:25:34.082935 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a is running failed: container process not found" containerID="c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 09:25:34 crc kubenswrapper[4744]: E1008 09:25:34.083237 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a is running failed: container process not found" containerID="c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 09:25:34 crc kubenswrapper[4744]: E1008 09:25:34.083464 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a is running failed: container process not found" containerID="c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 09:25:34 crc kubenswrapper[4744]: E1008 09:25:34.083496 4744 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-226vj" podUID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerName="registry-server" Oct 08 09:25:35 crc kubenswrapper[4744]: I1008 09:25:35.254232 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:35 crc kubenswrapper[4744]: I1008 09:25:35.321040 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dae8c09-ed33-493e-baad-730bd1cdc511-catalog-content\") pod \"8dae8c09-ed33-493e-baad-730bd1cdc511\" (UID: \"8dae8c09-ed33-493e-baad-730bd1cdc511\") " Oct 08 09:25:35 crc kubenswrapper[4744]: I1008 09:25:35.321116 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl59k\" (UniqueName: \"kubernetes.io/projected/8dae8c09-ed33-493e-baad-730bd1cdc511-kube-api-access-fl59k\") pod \"8dae8c09-ed33-493e-baad-730bd1cdc511\" (UID: \"8dae8c09-ed33-493e-baad-730bd1cdc511\") " Oct 08 09:25:35 crc kubenswrapper[4744]: I1008 09:25:35.351866 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dae8c09-ed33-493e-baad-730bd1cdc511-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8dae8c09-ed33-493e-baad-730bd1cdc511" (UID: "8dae8c09-ed33-493e-baad-730bd1cdc511"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:25:35 crc kubenswrapper[4744]: I1008 09:25:35.352412 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dae8c09-ed33-493e-baad-730bd1cdc511-kube-api-access-fl59k" (OuterVolumeSpecName: "kube-api-access-fl59k") pod "8dae8c09-ed33-493e-baad-730bd1cdc511" (UID: "8dae8c09-ed33-493e-baad-730bd1cdc511"). InnerVolumeSpecName "kube-api-access-fl59k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:25:35 crc kubenswrapper[4744]: I1008 09:25:35.426670 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dae8c09-ed33-493e-baad-730bd1cdc511-utilities\") pod \"8dae8c09-ed33-493e-baad-730bd1cdc511\" (UID: \"8dae8c09-ed33-493e-baad-730bd1cdc511\") " Oct 08 09:25:35 crc kubenswrapper[4744]: I1008 09:25:35.427249 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dae8c09-ed33-493e-baad-730bd1cdc511-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:25:35 crc kubenswrapper[4744]: I1008 09:25:35.427265 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl59k\" (UniqueName: \"kubernetes.io/projected/8dae8c09-ed33-493e-baad-730bd1cdc511-kube-api-access-fl59k\") on node \"crc\" DevicePath \"\"" Oct 08 09:25:35 crc kubenswrapper[4744]: I1008 09:25:35.431227 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dae8c09-ed33-493e-baad-730bd1cdc511-utilities" (OuterVolumeSpecName: "utilities") pod "8dae8c09-ed33-493e-baad-730bd1cdc511" (UID: "8dae8c09-ed33-493e-baad-730bd1cdc511"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:25:35 crc kubenswrapper[4744]: I1008 09:25:35.529686 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dae8c09-ed33-493e-baad-730bd1cdc511-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:25:35 crc kubenswrapper[4744]: E1008 09:25:35.679912 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr" podUID="58decb1a-987e-4366-9bf8-ad4bf73e5969" Oct 08 09:25:35 crc kubenswrapper[4744]: E1008 09:25:35.680603 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw" podUID="f3456365-6f0d-4087-b753-40292c80bf12" Oct 08 09:25:35 crc kubenswrapper[4744]: E1008 09:25:35.700013 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" podUID="b2bb8369-17dd-4f3b-8e8d-4af7895d892c" Oct 08 09:25:35 crc kubenswrapper[4744]: E1008 09:25:35.736926 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" podUID="5fae63cb-c599-447b-bafb-2ed6bad836d3" Oct 08 09:25:35 crc kubenswrapper[4744]: I1008 09:25:35.774897 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jczf2"] Oct 08 09:25:35 crc kubenswrapper[4744]: E1008 09:25:35.943077 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" podUID="48158343-3842-44df-86c8-3a2f0e06a09c" Oct 08 09:25:35 crc kubenswrapper[4744]: E1008 09:25:35.987248 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" podUID="622e796c-d598-4721-89bd-c0cb3f83a1ad" Oct 08 09:25:35 crc kubenswrapper[4744]: E1008 09:25:35.987831 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" podUID="18a79d6e-9afe-43b5-8cf1-b5a991d2fd49" Oct 08 09:25:36 crc kubenswrapper[4744]: E1008 09:25:36.021835 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" podUID="72a5635b-6920-4c2b-b91c-7621bc933959" Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.044289 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" event={"ID":"b2bb8369-17dd-4f3b-8e8d-4af7895d892c","Type":"ContainerStarted","Data":"e421dff60b49bcb1266a6e56a310dbb0f53bd14cdd6f4990e79b5448b9f63660"} Oct 08 09:25:36 crc kubenswrapper[4744]: E1008 09:25:36.051935 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" podUID="b2bb8369-17dd-4f3b-8e8d-4af7895d892c" Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.066171 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" event={"ID":"5fae63cb-c599-447b-bafb-2ed6bad836d3","Type":"ContainerStarted","Data":"376701c1add77fab4efdb18307dcc7dde735fe613c2b236edf8cb6c48fcf1a11"} Oct 08 09:25:36 crc kubenswrapper[4744]: E1008 09:25:36.069726 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" podUID="5fae63cb-c599-447b-bafb-2ed6bad836d3" Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.077595 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw" event={"ID":"f3456365-6f0d-4087-b753-40292c80bf12","Type":"ContainerStarted","Data":"01fb1578bdfd7efcbfacb7b3e42ec14e5fd7b4c89be10dc10610787c372c470d"} Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.118170 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" event={"ID":"72a5635b-6920-4c2b-b91c-7621bc933959","Type":"ContainerStarted","Data":"50615b081b016972b94949488e40613f63cfdf79f15c2b7baacb9df906a01791"} Oct 08 09:25:36 crc kubenswrapper[4744]: E1008 09:25:36.129731 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" podUID="72a5635b-6920-4c2b-b91c-7621bc933959" Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.131663 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" event={"ID":"622e796c-d598-4721-89bd-c0cb3f83a1ad","Type":"ContainerStarted","Data":"9f2af50de5ae1a1b6d2a3ead97c5b479542e320009a633a5813923478c0b6d9f"} Oct 08 09:25:36 crc kubenswrapper[4744]: E1008 09:25:36.133346 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:ec11cb8711bd1af22db3c84aa854349ee46191add3db45aecfabb1d8410c04d0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" podUID="622e796c-d598-4721-89bd-c0cb3f83a1ad" Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.135926 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczf2" event={"ID":"23976bfb-8e75-4752-99ef-250d8b0d1e9d","Type":"ContainerStarted","Data":"c06ff8eee70fe295174ab784e0f928ff24461b981f9420dab0da8b1868a45376"} Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.147438 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-226vj" event={"ID":"8dae8c09-ed33-493e-baad-730bd1cdc511","Type":"ContainerDied","Data":"982c891c6c804b1d8917002d365332ce1d23185b8bbf171b6401d69f40fde797"} Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.147502 4744 scope.go:117] "RemoveContainer" containerID="c41c7cf856ae21e77e3cc70857c19cbd5a3567bab7800d378836bac1774bd22a" Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.147646 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-226vj" Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.175179 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" event={"ID":"18a79d6e-9afe-43b5-8cf1-b5a991d2fd49","Type":"ContainerStarted","Data":"7ce9e3847b1db258a8686b7e2c39cab9234208643ddc2e43da809722546a593e"} Oct 08 09:25:36 crc kubenswrapper[4744]: E1008 09:25:36.197891 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:063a7e65b4ba98f0506f269ff7525b446eae06a5ed4a61c18ffa33a886500867\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" podUID="18a79d6e-9afe-43b5-8cf1-b5a991d2fd49" Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.210022 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" event={"ID":"48158343-3842-44df-86c8-3a2f0e06a09c","Type":"ContainerStarted","Data":"bb8466c04e203b95319b73ed287e5a22b5e5f0681157525ecd2fcc39057694d7"} Oct 08 09:25:36 crc kubenswrapper[4744]: E1008 09:25:36.220883 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" podUID="48158343-3842-44df-86c8-3a2f0e06a09c" Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.243558 4744 scope.go:117] "RemoveContainer" containerID="0cadae97da6e2218c0d41f376fc75918977e9300d9f3df63dfe57956d27b6c3a" Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.264689 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr" event={"ID":"58decb1a-987e-4366-9bf8-ad4bf73e5969","Type":"ContainerStarted","Data":"99d0c333aa35ddd5605ef48d4e7ea4185f63875ed019b3f022a33c51f90ae085"} Oct 08 09:25:36 crc kubenswrapper[4744]: E1008 09:25:36.265206 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" podUID="663cead5-7b5e-4940-94f9-5974f0424eda" Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.333536 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-226vj"] Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.335864 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-226vj"] Oct 08 09:25:36 crc kubenswrapper[4744]: I1008 09:25:36.451793 4744 scope.go:117] "RemoveContainer" containerID="27804c337514233ed1dd659773cff245ba0ec57691ca78b3851a4c9807d6470e" Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.319582 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb" event={"ID":"6126b107-8c51-4467-9c40-e2c4b268cec5","Type":"ContainerStarted","Data":"4f92b19d94944edeeaaf9b919eaa1274e8c103be59ed399cddec174251e3dd4d"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.336264 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x" event={"ID":"60f0c455-812d-40fe-b1b6-5aa75c34753c","Type":"ContainerStarted","Data":"81d1cf7e5a5c326feeefc0962e6088e8e34c547cdf9a8e6dac5913221b909501"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.385354 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2" event={"ID":"6d6f4234-2b91-4399-8aa2-f1064fde501c","Type":"ContainerStarted","Data":"f0fccd977b13e54bf61f8a258c84dff6a186a38e71a76459c42c03f3d73503ba"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.411918 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" event={"ID":"663cead5-7b5e-4940-94f9-5974f0424eda","Type":"ContainerStarted","Data":"e5082e46c7e6a1ee2b2c28b1cad7d878609b15952991b39147a24080453510ae"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.420636 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2" event={"ID":"3eeb379f-9005-4908-ac9e-6dfb06405fb2","Type":"ContainerStarted","Data":"24689f39021de5f5a6b3f041c91f4ff886d0b70845a1b0bbfe74cab90ba6615b"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.442647 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c" event={"ID":"2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34","Type":"ContainerStarted","Data":"29c8cf4b330b29f25ce4802c2e64f9f85aa301237527c411b1ca9fe55dbe96c1"} Oct 08 09:25:37 crc kubenswrapper[4744]: E1008 09:25:37.453059 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" podUID="663cead5-7b5e-4940-94f9-5974f0424eda" Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.530303 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dae8c09-ed33-493e-baad-730bd1cdc511" path="/var/lib/kubelet/pods/8dae8c09-ed33-493e-baad-730bd1cdc511/volumes" Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.547290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl" event={"ID":"254b5ec4-bc16-4e8a-a61b-8a4a588dd629","Type":"ContainerStarted","Data":"b2af00806c8c1940cd82890494504bd457383d36ac4999304a41de5268bd6950"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.547464 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4" event={"ID":"5bf79b31-d3bc-4422-9302-ac7a55a1d95d","Type":"ContainerStarted","Data":"465cedcbc4562678b6ebe730b9b6c447c4982eae464f7d183d100ef1d25faab5"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.562217 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb" event={"ID":"44cd0502-66fb-45ef-888a-da6273df7056","Type":"ContainerStarted","Data":"aba02b176da9e7ce30477f43c4ff7de215e4e3f357cc96f550441e0bf425cb0a"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.611447 4744 generic.go:334] "Generic (PLEG): container finished" podID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" containerID="3d481e70aab4facb58baefb548e0da09890a10a495e1a8ae6c736677d7708022" exitCode=0 Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.611853 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczf2" event={"ID":"23976bfb-8e75-4752-99ef-250d8b0d1e9d","Type":"ContainerDied","Data":"3d481e70aab4facb58baefb548e0da09890a10a495e1a8ae6c736677d7708022"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.629602 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ztld4" podStartSLOduration=4.860733824 podStartE2EDuration="29.629585715s" podCreationTimestamp="2025-10-08 09:25:08 +0000 UTC" firstStartedPulling="2025-10-08 09:25:10.572783549 +0000 UTC m=+805.820428788" lastFinishedPulling="2025-10-08 09:25:35.34163544 +0000 UTC m=+830.589280679" observedRunningTime="2025-10-08 09:25:37.627530726 +0000 UTC m=+832.875175975" watchObservedRunningTime="2025-10-08 09:25:37.629585715 +0000 UTC m=+832.877230954" Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.644729 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" event={"ID":"12b5913f-74df-4ba6-8aa8-af5989327ddd","Type":"ContainerStarted","Data":"7d4575c54ce205654e94dd16b7c9b59d917b4e6aaa820402841668637be04287"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.645669 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.663430 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc" event={"ID":"e991d608-9aa2-4028-8080-a6bd7dca66c8","Type":"ContainerStarted","Data":"6ef8b45f670ed7d1bfcfc03ff7059dcce812c690a7ce873ad62b2ee9a93dd16c"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.672995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv" event={"ID":"3a897464-5474-43db-99d5-35f691f64220","Type":"ContainerStarted","Data":"567a2877cd50601bfbbea35b0133df92e1ce7711988120313e05d810ac642746"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.680925 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" event={"ID":"c5096473-ff21-474e-b7fc-03196e4ee5a8","Type":"ContainerStarted","Data":"47281bb0bbbdd494e9e69aa43a441e4c92353a26129a992480ffa727303a7d64"} Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.686487 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb" event={"ID":"8a43b240-9fe3-4c10-afa9-7e78777e76d5","Type":"ContainerStarted","Data":"0198eb9035ca8da738f0d774c107d41fcf9c900573cfe33978f45d7e939b6dae"} Oct 08 09:25:37 crc kubenswrapper[4744]: E1008 09:25:37.692606 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" podUID="72a5635b-6920-4c2b-b91c-7621bc933959" Oct 08 09:25:37 crc kubenswrapper[4744]: E1008 09:25:37.692945 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:063a7e65b4ba98f0506f269ff7525b446eae06a5ed4a61c18ffa33a886500867\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" podUID="18a79d6e-9afe-43b5-8cf1-b5a991d2fd49" Oct 08 09:25:37 crc kubenswrapper[4744]: E1008 09:25:37.693000 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:ec11cb8711bd1af22db3c84aa854349ee46191add3db45aecfabb1d8410c04d0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" podUID="622e796c-d598-4721-89bd-c0cb3f83a1ad" Oct 08 09:25:37 crc kubenswrapper[4744]: E1008 09:25:37.693190 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" podUID="5fae63cb-c599-447b-bafb-2ed6bad836d3" Oct 08 09:25:37 crc kubenswrapper[4744]: E1008 09:25:37.703065 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" podUID="48158343-3842-44df-86c8-3a2f0e06a09c" Oct 08 09:25:37 crc kubenswrapper[4744]: I1008 09:25:37.757996 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" podStartSLOduration=4.897184496 podStartE2EDuration="29.757976861s" podCreationTimestamp="2025-10-08 09:25:08 +0000 UTC" firstStartedPulling="2025-10-08 09:25:10.503851382 +0000 UTC m=+805.751496621" lastFinishedPulling="2025-10-08 09:25:35.364643747 +0000 UTC m=+830.612288986" observedRunningTime="2025-10-08 09:25:37.747630465 +0000 UTC m=+832.995275704" watchObservedRunningTime="2025-10-08 09:25:37.757976861 +0000 UTC m=+833.005622100" Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.696088 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb" event={"ID":"44cd0502-66fb-45ef-888a-da6273df7056","Type":"ContainerStarted","Data":"1a6dda9f88bf9eee018e3e19c65262365c534bbaade50b2f5648fc8cceaa714c"} Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.696231 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb" Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.701118 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl" event={"ID":"254b5ec4-bc16-4e8a-a61b-8a4a588dd629","Type":"ContainerStarted","Data":"f0093bc4bc26d080de87b43e09381039ae68e5eb37042b9ebebd42d96cde97c6"} Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.701273 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl" Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.703872 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv" event={"ID":"3a897464-5474-43db-99d5-35f691f64220","Type":"ContainerStarted","Data":"5c4c352f928da8e87f0f831ce62dcb7914ef0b5e72e31a0e3bca3999d0f75225"} Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.706603 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb" event={"ID":"6126b107-8c51-4467-9c40-e2c4b268cec5","Type":"ContainerStarted","Data":"62afd59a2e2a3a97e607b1416453235cc5acea310180b5f5cde94bf6c7118555"} Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.706907 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb" Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.710162 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x" event={"ID":"60f0c455-812d-40fe-b1b6-5aa75c34753c","Type":"ContainerStarted","Data":"530e7b2b62aeb37fbdfda427a7995ca0a23ff7b98ea97ff6b939e51d41ca9edd"} Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.710206 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x" Oct 08 09:25:38 crc kubenswrapper[4744]: E1008 09:25:38.710785 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" podUID="663cead5-7b5e-4940-94f9-5974f0424eda" Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.760372 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb" podStartSLOduration=6.44243771 podStartE2EDuration="32.760345279s" podCreationTimestamp="2025-10-08 09:25:06 +0000 UTC" firstStartedPulling="2025-10-08 09:25:08.851319519 +0000 UTC m=+804.098964758" lastFinishedPulling="2025-10-08 09:25:35.169227088 +0000 UTC m=+830.416872327" observedRunningTime="2025-10-08 09:25:38.744458046 +0000 UTC m=+833.992103285" watchObservedRunningTime="2025-10-08 09:25:38.760345279 +0000 UTC m=+834.007990518" Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.846023 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl" podStartSLOduration=6.489413371 podStartE2EDuration="32.846000675s" podCreationTimestamp="2025-10-08 09:25:06 +0000 UTC" firstStartedPulling="2025-10-08 09:25:08.818122421 +0000 UTC m=+804.065767660" lastFinishedPulling="2025-10-08 09:25:35.174709725 +0000 UTC m=+830.422354964" observedRunningTime="2025-10-08 09:25:38.809900535 +0000 UTC m=+834.057545794" watchObservedRunningTime="2025-10-08 09:25:38.846000675 +0000 UTC m=+834.093645914" Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.846866 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x" podStartSLOduration=6.149873212 podStartE2EDuration="30.84685983s" podCreationTimestamp="2025-10-08 09:25:08 +0000 UTC" firstStartedPulling="2025-10-08 09:25:10.471854048 +0000 UTC m=+805.719499287" lastFinishedPulling="2025-10-08 09:25:35.168840666 +0000 UTC m=+830.416485905" observedRunningTime="2025-10-08 09:25:38.842080044 +0000 UTC m=+834.089725293" watchObservedRunningTime="2025-10-08 09:25:38.84685983 +0000 UTC m=+834.094505059" Oct 08 09:25:38 crc kubenswrapper[4744]: I1008 09:25:38.880314 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb" podStartSLOduration=6.3998378670000005 podStartE2EDuration="31.880278674s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:09.694934636 +0000 UTC m=+804.942579875" lastFinishedPulling="2025-10-08 09:25:35.175375443 +0000 UTC m=+830.423020682" observedRunningTime="2025-10-08 09:25:38.875714084 +0000 UTC m=+834.123359333" watchObservedRunningTime="2025-10-08 09:25:38.880278674 +0000 UTC m=+834.127923913" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.716548 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc" event={"ID":"e991d608-9aa2-4028-8080-a6bd7dca66c8","Type":"ContainerStarted","Data":"10bb581353c85d27515e4a3437c88d00258535e47091ca0bf1481df1cd8a7a49"} Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.716904 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.718048 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw" event={"ID":"f3456365-6f0d-4087-b753-40292c80bf12","Type":"ContainerStarted","Data":"3c44fc0307b9099f8cdbe4c033acc97b2b7bfaf2b8d74f8b6aed489a33efc029"} Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.718175 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.719930 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" event={"ID":"c5096473-ff21-474e-b7fc-03196e4ee5a8","Type":"ContainerStarted","Data":"f15f97ebb847848ea2d9fdd905fb327ec4ed9ceca278a46d5c079098a73328e6"} Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.720050 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.721771 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczf2" event={"ID":"23976bfb-8e75-4752-99ef-250d8b0d1e9d","Type":"ContainerStarted","Data":"30d8ebe2348497efe6f8e324956850fa87b6996ac75c373486e34b823d5205fd"} Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.723374 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c" event={"ID":"2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34","Type":"ContainerStarted","Data":"892ecbfe171093d5ce74e0179762e1a5cff5b1f5a3a2d79eaa71f46c57e2b749"} Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.723760 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.724928 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr" event={"ID":"58decb1a-987e-4366-9bf8-ad4bf73e5969","Type":"ContainerStarted","Data":"32b75376373e08d0f3e098dfe0babf5b45e4f0a1bee542399da2516fae8b3c2e"} Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.725331 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.726758 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb" event={"ID":"8a43b240-9fe3-4c10-afa9-7e78777e76d5","Type":"ContainerStarted","Data":"25267c8e22579005471cf7d1e9168c7212c5e68fe5b4f662a4126658d45f5342"} Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.727181 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.728429 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2" event={"ID":"3eeb379f-9005-4908-ac9e-6dfb06405fb2","Type":"ContainerStarted","Data":"f111213492dd7998e58e29f5585eb62ec80c40cd800b8e122b3d5eeafea2bf36"} Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.728833 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.730146 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2" event={"ID":"6d6f4234-2b91-4399-8aa2-f1064fde501c","Type":"ContainerStarted","Data":"4f4e2bc196e921deea9f72d6334ab460bb2f6c88fd3292ff3397e608fe5ae9e4"} Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.730576 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.732242 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" event={"ID":"b2bb8369-17dd-4f3b-8e8d-4af7895d892c","Type":"ContainerStarted","Data":"dfa66e6c49c3ec735d8e37160961962de2acad784dc13de309b5a79b70949d19"} Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.732608 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.733659 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.739292 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc" podStartSLOduration=7.902645755 podStartE2EDuration="32.73926912s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:10.337663637 +0000 UTC m=+805.585308876" lastFinishedPulling="2025-10-08 09:25:35.174287002 +0000 UTC m=+830.421932241" observedRunningTime="2025-10-08 09:25:39.73613162 +0000 UTC m=+834.983776879" watchObservedRunningTime="2025-10-08 09:25:39.73926912 +0000 UTC m=+834.986914359" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.794159 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr" podStartSLOduration=5.534795478 podStartE2EDuration="32.794138176s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:09.657906349 +0000 UTC m=+804.905551588" lastFinishedPulling="2025-10-08 09:25:36.917249047 +0000 UTC m=+832.164894286" observedRunningTime="2025-10-08 09:25:39.788107574 +0000 UTC m=+835.035752823" watchObservedRunningTime="2025-10-08 09:25:39.794138176 +0000 UTC m=+835.041783405" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.864458 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2" podStartSLOduration=8.163471491 podStartE2EDuration="33.864432164s" podCreationTimestamp="2025-10-08 09:25:06 +0000 UTC" firstStartedPulling="2025-10-08 09:25:09.473352469 +0000 UTC m=+804.720997708" lastFinishedPulling="2025-10-08 09:25:35.174313142 +0000 UTC m=+830.421958381" observedRunningTime="2025-10-08 09:25:39.86116672 +0000 UTC m=+835.108811969" watchObservedRunningTime="2025-10-08 09:25:39.864432164 +0000 UTC m=+835.112077403" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.889348 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv" podStartSLOduration=8.050778542 podStartE2EDuration="32.889318823s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:10.337209903 +0000 UTC m=+805.584855142" lastFinishedPulling="2025-10-08 09:25:35.175750184 +0000 UTC m=+830.423395423" observedRunningTime="2025-10-08 09:25:39.886431551 +0000 UTC m=+835.134076960" watchObservedRunningTime="2025-10-08 09:25:39.889318823 +0000 UTC m=+835.136964062" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.921157 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb" podStartSLOduration=7.24093149 podStartE2EDuration="33.921135502s" podCreationTimestamp="2025-10-08 09:25:06 +0000 UTC" firstStartedPulling="2025-10-08 09:25:08.49513435 +0000 UTC m=+803.742779589" lastFinishedPulling="2025-10-08 09:25:35.175338362 +0000 UTC m=+830.422983601" observedRunningTime="2025-10-08 09:25:39.913493774 +0000 UTC m=+835.161139033" watchObservedRunningTime="2025-10-08 09:25:39.921135502 +0000 UTC m=+835.168780741" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.971120 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" podStartSLOduration=8.625373069 podStartE2EDuration="32.971094979s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:10.847008679 +0000 UTC m=+806.094653918" lastFinishedPulling="2025-10-08 09:25:35.192730589 +0000 UTC m=+830.440375828" observedRunningTime="2025-10-08 09:25:39.958230661 +0000 UTC m=+835.205875910" watchObservedRunningTime="2025-10-08 09:25:39.971094979 +0000 UTC m=+835.218740218" Oct 08 09:25:39 crc kubenswrapper[4744]: I1008 09:25:39.992322 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c" podStartSLOduration=7.818702808 podStartE2EDuration="32.992294434s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:10.019110302 +0000 UTC m=+805.266755541" lastFinishedPulling="2025-10-08 09:25:35.192701928 +0000 UTC m=+830.440347167" observedRunningTime="2025-10-08 09:25:39.991883072 +0000 UTC m=+835.239528331" watchObservedRunningTime="2025-10-08 09:25:39.992294434 +0000 UTC m=+835.239939663" Oct 08 09:25:40 crc kubenswrapper[4744]: I1008 09:25:40.014407 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2" podStartSLOduration=8.159463378 podStartE2EDuration="33.014390275s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:10.337966256 +0000 UTC m=+805.585611495" lastFinishedPulling="2025-10-08 09:25:35.192893153 +0000 UTC m=+830.440538392" observedRunningTime="2025-10-08 09:25:40.011686768 +0000 UTC m=+835.259332027" watchObservedRunningTime="2025-10-08 09:25:40.014390275 +0000 UTC m=+835.262035514" Oct 08 09:25:40 crc kubenswrapper[4744]: I1008 09:25:40.036494 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" podStartSLOduration=4.405768871 podStartE2EDuration="33.036474685s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:10.197042231 +0000 UTC m=+805.444687470" lastFinishedPulling="2025-10-08 09:25:38.827748045 +0000 UTC m=+834.075393284" observedRunningTime="2025-10-08 09:25:40.0310381 +0000 UTC m=+835.278683359" watchObservedRunningTime="2025-10-08 09:25:40.036474685 +0000 UTC m=+835.284119924" Oct 08 09:25:40 crc kubenswrapper[4744]: I1008 09:25:40.053624 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw" podStartSLOduration=5.704141095 podStartE2EDuration="32.053602064s" podCreationTimestamp="2025-10-08 09:25:08 +0000 UTC" firstStartedPulling="2025-10-08 09:25:10.482822291 +0000 UTC m=+805.730467530" lastFinishedPulling="2025-10-08 09:25:36.83228326 +0000 UTC m=+832.079928499" observedRunningTime="2025-10-08 09:25:40.049501028 +0000 UTC m=+835.297146287" watchObservedRunningTime="2025-10-08 09:25:40.053602064 +0000 UTC m=+835.301247293" Oct 08 09:25:40 crc kubenswrapper[4744]: I1008 09:25:40.746121 4744 generic.go:334] "Generic (PLEG): container finished" podID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" containerID="30d8ebe2348497efe6f8e324956850fa87b6996ac75c373486e34b823d5205fd" exitCode=0 Oct 08 09:25:40 crc kubenswrapper[4744]: I1008 09:25:40.747734 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczf2" event={"ID":"23976bfb-8e75-4752-99ef-250d8b0d1e9d","Type":"ContainerDied","Data":"30d8ebe2348497efe6f8e324956850fa87b6996ac75c373486e34b823d5205fd"} Oct 08 09:25:41 crc kubenswrapper[4744]: I1008 09:25:41.758249 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczf2" event={"ID":"23976bfb-8e75-4752-99ef-250d8b0d1e9d","Type":"ContainerStarted","Data":"0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608"} Oct 08 09:25:41 crc kubenswrapper[4744]: I1008 09:25:41.762274 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5df598886f-5vz2c" Oct 08 09:25:41 crc kubenswrapper[4744]: I1008 09:25:41.762931 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-nzjx2" Oct 08 09:25:41 crc kubenswrapper[4744]: I1008 09:25:41.766174 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-csxmb" Oct 08 09:25:41 crc kubenswrapper[4744]: I1008 09:25:41.803887 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jczf2" podStartSLOduration=13.22480221 podStartE2EDuration="16.803856507s" podCreationTimestamp="2025-10-08 09:25:25 +0000 UTC" firstStartedPulling="2025-10-08 09:25:37.615420901 +0000 UTC m=+832.863066140" lastFinishedPulling="2025-10-08 09:25:41.194475188 +0000 UTC m=+836.442120437" observedRunningTime="2025-10-08 09:25:41.784872665 +0000 UTC m=+837.032517924" watchObservedRunningTime="2025-10-08 09:25:41.803856507 +0000 UTC m=+837.051501746" Oct 08 09:25:45 crc kubenswrapper[4744]: I1008 09:25:45.431150 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:45 crc kubenswrapper[4744]: I1008 09:25:45.431722 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:46 crc kubenswrapper[4744]: I1008 09:25:46.493792 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jczf2" podUID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" containerName="registry-server" probeResult="failure" output=< Oct 08 09:25:46 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 09:25:46 crc kubenswrapper[4744]: > Oct 08 09:25:47 crc kubenswrapper[4744]: I1008 09:25:47.221027 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-b9ntl" Oct 08 09:25:47 crc kubenswrapper[4744]: I1008 09:25:47.277519 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-f9zmb" Oct 08 09:25:47 crc kubenswrapper[4744]: I1008 09:25:47.762914 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-q9hcb" Oct 08 09:25:47 crc kubenswrapper[4744]: I1008 09:25:47.962044 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-lj9pr" Oct 08 09:25:48 crc kubenswrapper[4744]: I1008 09:25:48.189644 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-tbx59" Oct 08 09:25:48 crc kubenswrapper[4744]: I1008 09:25:48.468824 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-qppfc" Oct 08 09:25:48 crc kubenswrapper[4744]: I1008 09:25:48.635884 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-lbvtv" Oct 08 09:25:48 crc kubenswrapper[4744]: I1008 09:25:48.682709 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-crqx2" Oct 08 09:25:48 crc kubenswrapper[4744]: I1008 09:25:48.754144 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-gv4tw" Oct 08 09:25:49 crc kubenswrapper[4744]: I1008 09:25:49.188995 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-q884x" Oct 08 09:25:49 crc kubenswrapper[4744]: I1008 09:25:49.339556 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-56c698c775-bf9cm" Oct 08 09:25:49 crc kubenswrapper[4744]: I1008 09:25:49.820807 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" event={"ID":"48158343-3842-44df-86c8-3a2f0e06a09c","Type":"ContainerStarted","Data":"5e0fa37622af604bddaa73dc918abcee369ce8ffc4c50ceec4fe56f8580ddd2b"} Oct 08 09:25:49 crc kubenswrapper[4744]: I1008 09:25:49.821077 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" Oct 08 09:25:49 crc kubenswrapper[4744]: I1008 09:25:49.824002 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" event={"ID":"5fae63cb-c599-447b-bafb-2ed6bad836d3","Type":"ContainerStarted","Data":"d1ee3b905715c9a0e1064452d5884b564d8cc0265652c3260a14f6237d050435"} Oct 08 09:25:49 crc kubenswrapper[4744]: I1008 09:25:49.824226 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" Oct 08 09:25:49 crc kubenswrapper[4744]: I1008 09:25:49.839050 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" podStartSLOduration=3.92256771 podStartE2EDuration="42.839021546s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:09.983844885 +0000 UTC m=+805.231490124" lastFinishedPulling="2025-10-08 09:25:48.900298701 +0000 UTC m=+844.147943960" observedRunningTime="2025-10-08 09:25:49.838955974 +0000 UTC m=+845.086601223" watchObservedRunningTime="2025-10-08 09:25:49.839021546 +0000 UTC m=+845.086666785" Oct 08 09:25:49 crc kubenswrapper[4744]: I1008 09:25:49.866913 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" podStartSLOduration=3.837229989 podStartE2EDuration="42.866888183s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:09.99525474 +0000 UTC m=+805.242899979" lastFinishedPulling="2025-10-08 09:25:49.024912934 +0000 UTC m=+844.272558173" observedRunningTime="2025-10-08 09:25:49.865030729 +0000 UTC m=+845.112675988" watchObservedRunningTime="2025-10-08 09:25:49.866888183 +0000 UTC m=+845.114533442" Oct 08 09:25:50 crc kubenswrapper[4744]: I1008 09:25:50.004198 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-747747dfccwc6fw" Oct 08 09:25:52 crc kubenswrapper[4744]: I1008 09:25:52.851494 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" event={"ID":"72a5635b-6920-4c2b-b91c-7621bc933959","Type":"ContainerStarted","Data":"b0714d2e67067dd4aabd9c677209ed5e4f76e44839553d8b82373511712024f9"} Oct 08 09:25:52 crc kubenswrapper[4744]: I1008 09:25:52.852472 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" Oct 08 09:25:52 crc kubenswrapper[4744]: I1008 09:25:52.854161 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" event={"ID":"622e796c-d598-4721-89bd-c0cb3f83a1ad","Type":"ContainerStarted","Data":"8f292662e365041ad8e8abbddbf3e6248e5879c3e8aa46975995d4d30e2e2c94"} Oct 08 09:25:52 crc kubenswrapper[4744]: I1008 09:25:52.854585 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" Oct 08 09:25:52 crc kubenswrapper[4744]: I1008 09:25:52.859301 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" event={"ID":"18a79d6e-9afe-43b5-8cf1-b5a991d2fd49","Type":"ContainerStarted","Data":"b67799771b748bc2c0a56428c87560d229a19421748872ae7316c3007e8af2d6"} Oct 08 09:25:52 crc kubenswrapper[4744]: I1008 09:25:52.859596 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" Oct 08 09:25:52 crc kubenswrapper[4744]: I1008 09:25:52.879447 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" podStartSLOduration=3.440916537 podStartE2EDuration="45.87941323s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:09.485274079 +0000 UTC m=+804.732919318" lastFinishedPulling="2025-10-08 09:25:51.923770772 +0000 UTC m=+847.171416011" observedRunningTime="2025-10-08 09:25:52.873905494 +0000 UTC m=+848.121550743" watchObservedRunningTime="2025-10-08 09:25:52.87941323 +0000 UTC m=+848.127058489" Oct 08 09:25:52 crc kubenswrapper[4744]: I1008 09:25:52.905136 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" podStartSLOduration=4.072665283 podStartE2EDuration="46.905108276s" podCreationTimestamp="2025-10-08 09:25:06 +0000 UTC" firstStartedPulling="2025-10-08 09:25:09.166464087 +0000 UTC m=+804.414109316" lastFinishedPulling="2025-10-08 09:25:51.99890707 +0000 UTC m=+847.246552309" observedRunningTime="2025-10-08 09:25:52.898398033 +0000 UTC m=+848.146043302" watchObservedRunningTime="2025-10-08 09:25:52.905108276 +0000 UTC m=+848.152753515" Oct 08 09:25:52 crc kubenswrapper[4744]: I1008 09:25:52.922247 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" podStartSLOduration=3.303696235 podStartE2EDuration="45.922223765s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:09.239861163 +0000 UTC m=+804.487506402" lastFinishedPulling="2025-10-08 09:25:51.858388693 +0000 UTC m=+847.106033932" observedRunningTime="2025-10-08 09:25:52.917245402 +0000 UTC m=+848.164890661" watchObservedRunningTime="2025-10-08 09:25:52.922223765 +0000 UTC m=+848.169869004" Oct 08 09:25:53 crc kubenswrapper[4744]: I1008 09:25:53.870928 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" event={"ID":"663cead5-7b5e-4940-94f9-5974f0424eda","Type":"ContainerStarted","Data":"8fba1cd104b9ca5c393ac0a6f16afb1ca1452eec769f33db5c6956b73e9d46c8"} Oct 08 09:25:53 crc kubenswrapper[4744]: I1008 09:25:53.896101 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" podStartSLOduration=4.200482334 podStartE2EDuration="46.896082013s" podCreationTimestamp="2025-10-08 09:25:07 +0000 UTC" firstStartedPulling="2025-10-08 09:25:10.247234315 +0000 UTC m=+805.494879554" lastFinishedPulling="2025-10-08 09:25:52.942833994 +0000 UTC m=+848.190479233" observedRunningTime="2025-10-08 09:25:53.892432849 +0000 UTC m=+849.140078088" watchObservedRunningTime="2025-10-08 09:25:53.896082013 +0000 UTC m=+849.143727242" Oct 08 09:25:55 crc kubenswrapper[4744]: I1008 09:25:55.482309 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:55 crc kubenswrapper[4744]: I1008 09:25:55.545419 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:55 crc kubenswrapper[4744]: I1008 09:25:55.729631 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jczf2"] Oct 08 09:25:56 crc kubenswrapper[4744]: I1008 09:25:56.899552 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jczf2" podUID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" containerName="registry-server" containerID="cri-o://0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608" gracePeriod=2 Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.403099 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-m9prk" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.432141 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.469568 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-hsmrl" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.536297 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23976bfb-8e75-4752-99ef-250d8b0d1e9d-utilities\") pod \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\" (UID: \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\") " Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.536892 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23976bfb-8e75-4752-99ef-250d8b0d1e9d-catalog-content\") pod \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\" (UID: \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\") " Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.536970 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s4m9\" (UniqueName: \"kubernetes.io/projected/23976bfb-8e75-4752-99ef-250d8b0d1e9d-kube-api-access-7s4m9\") pod \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\" (UID: \"23976bfb-8e75-4752-99ef-250d8b0d1e9d\") " Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.537053 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23976bfb-8e75-4752-99ef-250d8b0d1e9d-utilities" (OuterVolumeSpecName: "utilities") pod "23976bfb-8e75-4752-99ef-250d8b0d1e9d" (UID: "23976bfb-8e75-4752-99ef-250d8b0d1e9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.537472 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23976bfb-8e75-4752-99ef-250d8b0d1e9d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.546681 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23976bfb-8e75-4752-99ef-250d8b0d1e9d-kube-api-access-7s4m9" (OuterVolumeSpecName: "kube-api-access-7s4m9") pod "23976bfb-8e75-4752-99ef-250d8b0d1e9d" (UID: "23976bfb-8e75-4752-99ef-250d8b0d1e9d"). InnerVolumeSpecName "kube-api-access-7s4m9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.585084 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23976bfb-8e75-4752-99ef-250d8b0d1e9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23976bfb-8e75-4752-99ef-250d8b0d1e9d" (UID: "23976bfb-8e75-4752-99ef-250d8b0d1e9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.638762 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23976bfb-8e75-4752-99ef-250d8b0d1e9d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.638793 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s4m9\" (UniqueName: \"kubernetes.io/projected/23976bfb-8e75-4752-99ef-250d8b0d1e9d-kube-api-access-7s4m9\") on node \"crc\" DevicePath \"\"" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.865067 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-xjxx8" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.909323 4744 generic.go:334] "Generic (PLEG): container finished" podID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" containerID="0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608" exitCode=0 Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.909392 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczf2" event={"ID":"23976bfb-8e75-4752-99ef-250d8b0d1e9d","Type":"ContainerDied","Data":"0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608"} Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.909413 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jczf2" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.909427 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jczf2" event={"ID":"23976bfb-8e75-4752-99ef-250d8b0d1e9d","Type":"ContainerDied","Data":"c06ff8eee70fe295174ab784e0f928ff24461b981f9420dab0da8b1868a45376"} Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.909446 4744 scope.go:117] "RemoveContainer" containerID="0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.932574 4744 scope.go:117] "RemoveContainer" containerID="30d8ebe2348497efe6f8e324956850fa87b6996ac75c373486e34b823d5205fd" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.952484 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jczf2"] Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.959035 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jczf2"] Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.967074 4744 scope.go:117] "RemoveContainer" containerID="3d481e70aab4facb58baefb548e0da09890a10a495e1a8ae6c736677d7708022" Oct 08 09:25:57 crc kubenswrapper[4744]: I1008 09:25:57.999671 4744 scope.go:117] "RemoveContainer" containerID="0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608" Oct 08 09:25:58 crc kubenswrapper[4744]: E1008 09:25:58.000203 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608\": container with ID starting with 0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608 not found: ID does not exist" containerID="0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608" Oct 08 09:25:58 crc kubenswrapper[4744]: I1008 09:25:58.000261 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608"} err="failed to get container status \"0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608\": rpc error: code = NotFound desc = could not find container \"0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608\": container with ID starting with 0e9819a1b14350c5e6cbee35286499643caf6b2490b38e2c55d3ebc4a0f48608 not found: ID does not exist" Oct 08 09:25:58 crc kubenswrapper[4744]: I1008 09:25:58.000299 4744 scope.go:117] "RemoveContainer" containerID="30d8ebe2348497efe6f8e324956850fa87b6996ac75c373486e34b823d5205fd" Oct 08 09:25:58 crc kubenswrapper[4744]: E1008 09:25:58.000764 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30d8ebe2348497efe6f8e324956850fa87b6996ac75c373486e34b823d5205fd\": container with ID starting with 30d8ebe2348497efe6f8e324956850fa87b6996ac75c373486e34b823d5205fd not found: ID does not exist" containerID="30d8ebe2348497efe6f8e324956850fa87b6996ac75c373486e34b823d5205fd" Oct 08 09:25:58 crc kubenswrapper[4744]: I1008 09:25:58.000803 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30d8ebe2348497efe6f8e324956850fa87b6996ac75c373486e34b823d5205fd"} err="failed to get container status \"30d8ebe2348497efe6f8e324956850fa87b6996ac75c373486e34b823d5205fd\": rpc error: code = NotFound desc = could not find container \"30d8ebe2348497efe6f8e324956850fa87b6996ac75c373486e34b823d5205fd\": container with ID starting with 30d8ebe2348497efe6f8e324956850fa87b6996ac75c373486e34b823d5205fd not found: ID does not exist" Oct 08 09:25:58 crc kubenswrapper[4744]: I1008 09:25:58.000834 4744 scope.go:117] "RemoveContainer" containerID="3d481e70aab4facb58baefb548e0da09890a10a495e1a8ae6c736677d7708022" Oct 08 09:25:58 crc kubenswrapper[4744]: E1008 09:25:58.001104 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d481e70aab4facb58baefb548e0da09890a10a495e1a8ae6c736677d7708022\": container with ID starting with 3d481e70aab4facb58baefb548e0da09890a10a495e1a8ae6c736677d7708022 not found: ID does not exist" containerID="3d481e70aab4facb58baefb548e0da09890a10a495e1a8ae6c736677d7708022" Oct 08 09:25:58 crc kubenswrapper[4744]: I1008 09:25:58.001124 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d481e70aab4facb58baefb548e0da09890a10a495e1a8ae6c736677d7708022"} err="failed to get container status \"3d481e70aab4facb58baefb548e0da09890a10a495e1a8ae6c736677d7708022\": rpc error: code = NotFound desc = could not find container \"3d481e70aab4facb58baefb548e0da09890a10a495e1a8ae6c736677d7708022\": container with ID starting with 3d481e70aab4facb58baefb548e0da09890a10a495e1a8ae6c736677d7708022 not found: ID does not exist" Oct 08 09:25:58 crc kubenswrapper[4744]: I1008 09:25:58.181826 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-xl4hw" Oct 08 09:25:58 crc kubenswrapper[4744]: I1008 09:25:58.374395 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-4xvzd" Oct 08 09:25:58 crc kubenswrapper[4744]: I1008 09:25:58.632014 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" Oct 08 09:25:58 crc kubenswrapper[4744]: I1008 09:25:58.634433 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-frfr7" Oct 08 09:25:59 crc kubenswrapper[4744]: I1008 09:25:59.462097 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" path="/var/lib/kubelet/pods/23976bfb-8e75-4752-99ef-250d8b0d1e9d/volumes" Oct 08 09:26:19 crc kubenswrapper[4744]: I1008 09:26:19.690476 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:26:19 crc kubenswrapper[4744]: I1008 09:26:19.690970 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.297058 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-9zj26"] Oct 08 09:26:21 crc kubenswrapper[4744]: E1008 09:26:21.297758 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerName="extract-content" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.297778 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerName="extract-content" Oct 08 09:26:21 crc kubenswrapper[4744]: E1008 09:26:21.297819 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" containerName="extract-utilities" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.297829 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" containerName="extract-utilities" Oct 08 09:26:21 crc kubenswrapper[4744]: E1008 09:26:21.297854 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" containerName="registry-server" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.297861 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" containerName="registry-server" Oct 08 09:26:21 crc kubenswrapper[4744]: E1008 09:26:21.297871 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerName="extract-utilities" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.297878 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerName="extract-utilities" Oct 08 09:26:21 crc kubenswrapper[4744]: E1008 09:26:21.297892 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerName="registry-server" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.297900 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerName="registry-server" Oct 08 09:26:21 crc kubenswrapper[4744]: E1008 09:26:21.297929 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" containerName="extract-content" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.297937 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" containerName="extract-content" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.298114 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="23976bfb-8e75-4752-99ef-250d8b0d1e9d" containerName="registry-server" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.298147 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dae8c09-ed33-493e-baad-730bd1cdc511" containerName="registry-server" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.299016 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.316645 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.316737 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-4qh2v" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.316906 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.323806 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.335736 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-9zj26"] Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.402987 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-82cxq"] Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.404276 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.409452 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.425520 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca76c78-de7e-4eb0-8059-5a1fbcd99796-config\") pod \"dnsmasq-dns-7bfcb9d745-9zj26\" (UID: \"eca76c78-de7e-4eb0-8059-5a1fbcd99796\") " pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.425572 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6f9w\" (UniqueName: \"kubernetes.io/projected/eca76c78-de7e-4eb0-8059-5a1fbcd99796-kube-api-access-v6f9w\") pod \"dnsmasq-dns-7bfcb9d745-9zj26\" (UID: \"eca76c78-de7e-4eb0-8059-5a1fbcd99796\") " pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.477755 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-82cxq"] Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.526816 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca76c78-de7e-4eb0-8059-5a1fbcd99796-config\") pod \"dnsmasq-dns-7bfcb9d745-9zj26\" (UID: \"eca76c78-de7e-4eb0-8059-5a1fbcd99796\") " pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.526876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6f9w\" (UniqueName: \"kubernetes.io/projected/eca76c78-de7e-4eb0-8059-5a1fbcd99796-kube-api-access-v6f9w\") pod \"dnsmasq-dns-7bfcb9d745-9zj26\" (UID: \"eca76c78-de7e-4eb0-8059-5a1fbcd99796\") " pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.526902 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-config\") pod \"dnsmasq-dns-758b79db4c-82cxq\" (UID: \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\") " pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.527012 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqw9d\" (UniqueName: \"kubernetes.io/projected/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-kube-api-access-tqw9d\") pod \"dnsmasq-dns-758b79db4c-82cxq\" (UID: \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\") " pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.527043 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-dns-svc\") pod \"dnsmasq-dns-758b79db4c-82cxq\" (UID: \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\") " pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.528176 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca76c78-de7e-4eb0-8059-5a1fbcd99796-config\") pod \"dnsmasq-dns-7bfcb9d745-9zj26\" (UID: \"eca76c78-de7e-4eb0-8059-5a1fbcd99796\") " pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.563989 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6f9w\" (UniqueName: \"kubernetes.io/projected/eca76c78-de7e-4eb0-8059-5a1fbcd99796-kube-api-access-v6f9w\") pod \"dnsmasq-dns-7bfcb9d745-9zj26\" (UID: \"eca76c78-de7e-4eb0-8059-5a1fbcd99796\") " pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.620633 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.627874 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqw9d\" (UniqueName: \"kubernetes.io/projected/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-kube-api-access-tqw9d\") pod \"dnsmasq-dns-758b79db4c-82cxq\" (UID: \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\") " pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.628774 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-dns-svc\") pod \"dnsmasq-dns-758b79db4c-82cxq\" (UID: \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\") " pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.629165 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-config\") pod \"dnsmasq-dns-758b79db4c-82cxq\" (UID: \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\") " pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.629863 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-dns-svc\") pod \"dnsmasq-dns-758b79db4c-82cxq\" (UID: \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\") " pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.629912 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-config\") pod \"dnsmasq-dns-758b79db4c-82cxq\" (UID: \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\") " pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.651933 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqw9d\" (UniqueName: \"kubernetes.io/projected/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-kube-api-access-tqw9d\") pod \"dnsmasq-dns-758b79db4c-82cxq\" (UID: \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\") " pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:21 crc kubenswrapper[4744]: I1008 09:26:21.720884 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:22 crc kubenswrapper[4744]: I1008 09:26:22.072315 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-82cxq"] Oct 08 09:26:22 crc kubenswrapper[4744]: I1008 09:26:22.083575 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 09:26:22 crc kubenswrapper[4744]: I1008 09:26:22.101078 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-9zj26"] Oct 08 09:26:22 crc kubenswrapper[4744]: I1008 09:26:22.101422 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-82cxq" event={"ID":"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011","Type":"ContainerStarted","Data":"52e7ada6a64094c67db720dda8c6e21d1a26de7353fee919ca06c40e15f1f90c"} Oct 08 09:26:22 crc kubenswrapper[4744]: W1008 09:26:22.115120 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeca76c78_de7e_4eb0_8059_5a1fbcd99796.slice/crio-0a1cc0bb614874c8d23e9493b3c2497f68e4f69243365ff6863ed311813231d3 WatchSource:0}: Error finding container 0a1cc0bb614874c8d23e9493b3c2497f68e4f69243365ff6863ed311813231d3: Status 404 returned error can't find the container with id 0a1cc0bb614874c8d23e9493b3c2497f68e4f69243365ff6863ed311813231d3 Oct 08 09:26:23 crc kubenswrapper[4744]: I1008 09:26:23.110987 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" event={"ID":"eca76c78-de7e-4eb0-8059-5a1fbcd99796","Type":"ContainerStarted","Data":"0a1cc0bb614874c8d23e9493b3c2497f68e4f69243365ff6863ed311813231d3"} Oct 08 09:26:23 crc kubenswrapper[4744]: I1008 09:26:23.988142 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-82cxq"] Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.034622 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-644597f84c-tcwbc"] Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.045832 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.051678 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-tcwbc"] Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.087855 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/683a64cc-3b26-482b-a1bd-9eac5f056a4b-config\") pod \"dnsmasq-dns-644597f84c-tcwbc\" (UID: \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\") " pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.087922 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/683a64cc-3b26-482b-a1bd-9eac5f056a4b-dns-svc\") pod \"dnsmasq-dns-644597f84c-tcwbc\" (UID: \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\") " pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.087988 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5ktb\" (UniqueName: \"kubernetes.io/projected/683a64cc-3b26-482b-a1bd-9eac5f056a4b-kube-api-access-l5ktb\") pod \"dnsmasq-dns-644597f84c-tcwbc\" (UID: \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\") " pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.197403 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/683a64cc-3b26-482b-a1bd-9eac5f056a4b-config\") pod \"dnsmasq-dns-644597f84c-tcwbc\" (UID: \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\") " pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.197468 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/683a64cc-3b26-482b-a1bd-9eac5f056a4b-dns-svc\") pod \"dnsmasq-dns-644597f84c-tcwbc\" (UID: \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\") " pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.197523 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5ktb\" (UniqueName: \"kubernetes.io/projected/683a64cc-3b26-482b-a1bd-9eac5f056a4b-kube-api-access-l5ktb\") pod \"dnsmasq-dns-644597f84c-tcwbc\" (UID: \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\") " pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.198539 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/683a64cc-3b26-482b-a1bd-9eac5f056a4b-dns-svc\") pod \"dnsmasq-dns-644597f84c-tcwbc\" (UID: \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\") " pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.198544 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/683a64cc-3b26-482b-a1bd-9eac5f056a4b-config\") pod \"dnsmasq-dns-644597f84c-tcwbc\" (UID: \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\") " pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.245322 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5ktb\" (UniqueName: \"kubernetes.io/projected/683a64cc-3b26-482b-a1bd-9eac5f056a4b-kube-api-access-l5ktb\") pod \"dnsmasq-dns-644597f84c-tcwbc\" (UID: \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\") " pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.405712 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.405783 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-9zj26"] Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.457693 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77597f887-ktnfw"] Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.459028 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.482814 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-ktnfw"] Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.605954 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-config\") pod \"dnsmasq-dns-77597f887-ktnfw\" (UID: \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\") " pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.606632 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shb5b\" (UniqueName: \"kubernetes.io/projected/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-kube-api-access-shb5b\") pod \"dnsmasq-dns-77597f887-ktnfw\" (UID: \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\") " pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.606662 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-dns-svc\") pod \"dnsmasq-dns-77597f887-ktnfw\" (UID: \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\") " pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.708771 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shb5b\" (UniqueName: \"kubernetes.io/projected/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-kube-api-access-shb5b\") pod \"dnsmasq-dns-77597f887-ktnfw\" (UID: \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\") " pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.708822 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-dns-svc\") pod \"dnsmasq-dns-77597f887-ktnfw\" (UID: \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\") " pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.708902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-config\") pod \"dnsmasq-dns-77597f887-ktnfw\" (UID: \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\") " pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.710152 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-dns-svc\") pod \"dnsmasq-dns-77597f887-ktnfw\" (UID: \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\") " pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.711758 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-config\") pod \"dnsmasq-dns-77597f887-ktnfw\" (UID: \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\") " pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.762547 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shb5b\" (UniqueName: \"kubernetes.io/projected/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-kube-api-access-shb5b\") pod \"dnsmasq-dns-77597f887-ktnfw\" (UID: \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\") " pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:24 crc kubenswrapper[4744]: I1008 09:26:24.847628 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.155189 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-tcwbc"] Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.238811 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.246443 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.264964 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.265058 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5qvmw" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.265401 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.265565 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.267340 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.275780 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.275842 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.295438 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.461548 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.461599 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6a6f5a1b-6abb-4568-89b5-a255f919cebb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.461616 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-config-data\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.461635 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.461672 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.461692 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.461710 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.461742 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.461778 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbscx\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-kube-api-access-lbscx\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.461806 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.461823 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6a6f5a1b-6abb-4568-89b5-a255f919cebb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.563035 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.563084 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.563109 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.563144 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.563204 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbscx\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-kube-api-access-lbscx\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.563237 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.563261 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6a6f5a1b-6abb-4568-89b5-a255f919cebb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.563300 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.563317 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6a6f5a1b-6abb-4568-89b5-a255f919cebb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.563347 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-config-data\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.563383 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.565798 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.568356 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-config-data\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.568456 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.568704 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.569080 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.570046 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.571109 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6a6f5a1b-6abb-4568-89b5-a255f919cebb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.578746 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6a6f5a1b-6abb-4568-89b5-a255f919cebb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.593732 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbscx\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-kube-api-access-lbscx\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.594331 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.594480 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.617529 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.657212 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.681536 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.682962 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.687983 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.688066 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.688192 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.688224 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.688064 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.688326 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.688237 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5sp9s" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.706619 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.768116 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.768179 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.768206 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.768228 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.768276 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dql57\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-kube-api-access-dql57\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.768304 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.768337 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.768362 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.768422 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.768444 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.768468 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.808478 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-ktnfw"] Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.870537 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.870584 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.870611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.870674 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dql57\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-kube-api-access-dql57\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.870709 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.870746 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.870775 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.870803 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.870833 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.870860 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.870930 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.871576 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.871601 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.871712 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.872323 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.874163 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.876161 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.883662 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.890675 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.899103 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dql57\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-kube-api-access-dql57\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.900109 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.911572 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:25 crc kubenswrapper[4744]: I1008 09:26:25.967574 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:26 crc kubenswrapper[4744]: W1008 09:26:26.016510 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00e8ebd9_71ce_42b9_8bd3_148d3f9309b9.slice/crio-0d321790518be440628efc9dae8c456dc785962d013bd221a6e93963d318f4dd WatchSource:0}: Error finding container 0d321790518be440628efc9dae8c456dc785962d013bd221a6e93963d318f4dd: Status 404 returned error can't find the container with id 0d321790518be440628efc9dae8c456dc785962d013bd221a6e93963d318f4dd Oct 08 09:26:26 crc kubenswrapper[4744]: I1008 09:26:26.068991 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:26:26 crc kubenswrapper[4744]: I1008 09:26:26.184090 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-ktnfw" event={"ID":"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9","Type":"ContainerStarted","Data":"0d321790518be440628efc9dae8c456dc785962d013bd221a6e93963d318f4dd"} Oct 08 09:26:26 crc kubenswrapper[4744]: I1008 09:26:26.220267 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-tcwbc" event={"ID":"683a64cc-3b26-482b-a1bd-9eac5f056a4b","Type":"ContainerStarted","Data":"6f565aaf3ce43663611692c6974a8667b8c0613d4a252797c2b93c29575c61bf"} Oct 08 09:26:26 crc kubenswrapper[4744]: I1008 09:26:26.759269 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 09:26:26 crc kubenswrapper[4744]: W1008 09:26:26.801328 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a6f5a1b_6abb_4568_89b5_a255f919cebb.slice/crio-eeedf4ac70685ed31ef68628b235f1aca550e7054155460c642b17603761d187 WatchSource:0}: Error finding container eeedf4ac70685ed31ef68628b235f1aca550e7054155460c642b17603761d187: Status 404 returned error can't find the container with id eeedf4ac70685ed31ef68628b235f1aca550e7054155460c642b17603761d187 Oct 08 09:26:26 crc kubenswrapper[4744]: I1008 09:26:26.985130 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 09:26:27 crc kubenswrapper[4744]: I1008 09:26:27.258939 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6a6f5a1b-6abb-4568-89b5-a255f919cebb","Type":"ContainerStarted","Data":"eeedf4ac70685ed31ef68628b235f1aca550e7054155460c642b17603761d187"} Oct 08 09:26:27 crc kubenswrapper[4744]: I1008 09:26:27.263999 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6","Type":"ContainerStarted","Data":"1df0e399ccf38a5de5ca1c8c40ff27077c3c6be4db4c989dbfa297e8dc1d8dab"} Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.016796 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.031317 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.033517 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.047743 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.048579 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.057247 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-jd28j" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.058142 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.061430 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.082362 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.167257 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d1412b08-0e26-455c-afb5-3f51b2cb3012-kolla-config\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.169249 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pmmn\" (UniqueName: \"kubernetes.io/projected/d1412b08-0e26-455c-afb5-3f51b2cb3012-kube-api-access-2pmmn\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.169528 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1412b08-0e26-455c-afb5-3f51b2cb3012-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.169673 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.170494 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d1412b08-0e26-455c-afb5-3f51b2cb3012-secrets\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.170639 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d1412b08-0e26-455c-afb5-3f51b2cb3012-config-data-default\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.170701 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1412b08-0e26-455c-afb5-3f51b2cb3012-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.170724 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d1412b08-0e26-455c-afb5-3f51b2cb3012-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.170809 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1412b08-0e26-455c-afb5-3f51b2cb3012-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.272553 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.272626 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d1412b08-0e26-455c-afb5-3f51b2cb3012-secrets\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.272658 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d1412b08-0e26-455c-afb5-3f51b2cb3012-config-data-default\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.272681 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1412b08-0e26-455c-afb5-3f51b2cb3012-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.272702 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d1412b08-0e26-455c-afb5-3f51b2cb3012-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.272721 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1412b08-0e26-455c-afb5-3f51b2cb3012-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.272746 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d1412b08-0e26-455c-afb5-3f51b2cb3012-kolla-config\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.272773 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pmmn\" (UniqueName: \"kubernetes.io/projected/d1412b08-0e26-455c-afb5-3f51b2cb3012-kube-api-access-2pmmn\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.272792 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1412b08-0e26-455c-afb5-3f51b2cb3012-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.282762 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d1412b08-0e26-455c-afb5-3f51b2cb3012-config-data-default\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.282766 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.283303 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1412b08-0e26-455c-afb5-3f51b2cb3012-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.283824 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1412b08-0e26-455c-afb5-3f51b2cb3012-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.283951 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d1412b08-0e26-455c-afb5-3f51b2cb3012-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.284274 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d1412b08-0e26-455c-afb5-3f51b2cb3012-kolla-config\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.297227 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1412b08-0e26-455c-afb5-3f51b2cb3012-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.299980 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d1412b08-0e26-455c-afb5-3f51b2cb3012-secrets\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.310172 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pmmn\" (UniqueName: \"kubernetes.io/projected/d1412b08-0e26-455c-afb5-3f51b2cb3012-kube-api-access-2pmmn\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.390009 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"d1412b08-0e26-455c-afb5-3f51b2cb3012\") " pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.503070 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.504506 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.523684 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-qhvvh" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.524189 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.524511 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.524516 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.525399 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.582479 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.582590 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.582664 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.582739 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.582772 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.582846 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.582921 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.582976 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ndnv\" (UniqueName: \"kubernetes.io/projected/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-kube-api-access-2ndnv\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.583019 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.675942 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.687120 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.687162 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ndnv\" (UniqueName: \"kubernetes.io/projected/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-kube-api-access-2ndnv\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.687193 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.687244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.687268 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.687302 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.687343 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.687365 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.687413 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.689572 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.690055 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.691023 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.691219 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.691598 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.703982 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.716970 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.723277 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.754808 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.756188 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.774315 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.775000 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-cwl7s" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.775142 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.807096 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.813523 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.814318 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ndnv\" (UniqueName: \"kubernetes.io/projected/ef3dc460-b269-4e4c-b1db-bfb550b58a7b-kube-api-access-2ndnv\") pod \"openstack-cell1-galera-0\" (UID: \"ef3dc460-b269-4e4c-b1db-bfb550b58a7b\") " pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.862754 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.892788 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8e37bd-80bc-4e95-8baf-f571adecb232-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.893217 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c8e37bd-80bc-4e95-8baf-f571adecb232-config-data\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.893452 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c8e37bd-80bc-4e95-8baf-f571adecb232-kolla-config\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.893553 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c8e37bd-80bc-4e95-8baf-f571adecb232-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.893725 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b75c\" (UniqueName: \"kubernetes.io/projected/4c8e37bd-80bc-4e95-8baf-f571adecb232-kube-api-access-8b75c\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.995548 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c8e37bd-80bc-4e95-8baf-f571adecb232-kolla-config\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.995605 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c8e37bd-80bc-4e95-8baf-f571adecb232-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.995626 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b75c\" (UniqueName: \"kubernetes.io/projected/4c8e37bd-80bc-4e95-8baf-f571adecb232-kube-api-access-8b75c\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.995685 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8e37bd-80bc-4e95-8baf-f571adecb232-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.995704 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c8e37bd-80bc-4e95-8baf-f571adecb232-config-data\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.996585 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c8e37bd-80bc-4e95-8baf-f571adecb232-config-data\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:28 crc kubenswrapper[4744]: I1008 09:26:28.996675 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c8e37bd-80bc-4e95-8baf-f571adecb232-kolla-config\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:29 crc kubenswrapper[4744]: I1008 09:26:29.004841 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8e37bd-80bc-4e95-8baf-f571adecb232-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:29 crc kubenswrapper[4744]: I1008 09:26:29.012109 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c8e37bd-80bc-4e95-8baf-f571adecb232-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:29 crc kubenswrapper[4744]: I1008 09:26:29.030585 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b75c\" (UniqueName: \"kubernetes.io/projected/4c8e37bd-80bc-4e95-8baf-f571adecb232-kube-api-access-8b75c\") pod \"memcached-0\" (UID: \"4c8e37bd-80bc-4e95-8baf-f571adecb232\") " pod="openstack/memcached-0" Oct 08 09:26:29 crc kubenswrapper[4744]: I1008 09:26:29.096857 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 09:26:29 crc kubenswrapper[4744]: I1008 09:26:29.850664 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 09:26:29 crc kubenswrapper[4744]: W1008 09:26:29.915420 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1412b08_0e26_455c_afb5_3f51b2cb3012.slice/crio-b32bb217f83c517477a6c062314a6678432de9a5688415eecf772e50f169178e WatchSource:0}: Error finding container b32bb217f83c517477a6c062314a6678432de9a5688415eecf772e50f169178e: Status 404 returned error can't find the container with id b32bb217f83c517477a6c062314a6678432de9a5688415eecf772e50f169178e Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.033134 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.039247 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.412995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ef3dc460-b269-4e4c-b1db-bfb550b58a7b","Type":"ContainerStarted","Data":"cbb219723ddf249ac819381e819565d807df7e93ba198ed971c0f6ecf3d79703"} Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.418227 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4c8e37bd-80bc-4e95-8baf-f571adecb232","Type":"ContainerStarted","Data":"acef2a7c79b8a2700cc1f705c793a48d31c0367f9a92d5cf47e5e3d641498c8a"} Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.424618 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d1412b08-0e26-455c-afb5-3f51b2cb3012","Type":"ContainerStarted","Data":"b32bb217f83c517477a6c062314a6678432de9a5688415eecf772e50f169178e"} Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.523745 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.525044 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.530489 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-7h88j" Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.554697 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.659956 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcm2r\" (UniqueName: \"kubernetes.io/projected/7eb871ad-c10b-48ea-b27e-f9dfca634845-kube-api-access-pcm2r\") pod \"kube-state-metrics-0\" (UID: \"7eb871ad-c10b-48ea-b27e-f9dfca634845\") " pod="openstack/kube-state-metrics-0" Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.761664 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcm2r\" (UniqueName: \"kubernetes.io/projected/7eb871ad-c10b-48ea-b27e-f9dfca634845-kube-api-access-pcm2r\") pod \"kube-state-metrics-0\" (UID: \"7eb871ad-c10b-48ea-b27e-f9dfca634845\") " pod="openstack/kube-state-metrics-0" Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.806725 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcm2r\" (UniqueName: \"kubernetes.io/projected/7eb871ad-c10b-48ea-b27e-f9dfca634845-kube-api-access-pcm2r\") pod \"kube-state-metrics-0\" (UID: \"7eb871ad-c10b-48ea-b27e-f9dfca634845\") " pod="openstack/kube-state-metrics-0" Oct 08 09:26:30 crc kubenswrapper[4744]: I1008 09:26:30.853742 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 09:26:31 crc kubenswrapper[4744]: I1008 09:26:31.541721 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 09:26:32 crc kubenswrapper[4744]: I1008 09:26:32.469192 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7eb871ad-c10b-48ea-b27e-f9dfca634845","Type":"ContainerStarted","Data":"a6feff7f60e80b53d447194024690ed34e47c66c7fd7be0df08ede3091ae3754"} Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.173030 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.174497 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.178460 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.178833 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-6nchw" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.179209 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.179328 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.202995 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.214500 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.336193 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9st4n\" (UniqueName: \"kubernetes.io/projected/a38f1196-55fe-462c-b012-be0d5fbe33ee-kube-api-access-9st4n\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.336787 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.336849 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a38f1196-55fe-462c-b012-be0d5fbe33ee-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.336907 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a38f1196-55fe-462c-b012-be0d5fbe33ee-config\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.336924 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a38f1196-55fe-462c-b012-be0d5fbe33ee-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.337108 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a38f1196-55fe-462c-b012-be0d5fbe33ee-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.337151 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a38f1196-55fe-462c-b012-be0d5fbe33ee-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.337195 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a38f1196-55fe-462c-b012-be0d5fbe33ee-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.396321 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-ff8dm"] Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.398047 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.401764 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-rtz5x" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.405323 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.408956 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-mhqr7"] Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.410137 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.414660 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.433393 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-ff8dm"] Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.441222 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mhqr7"] Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.441852 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a38f1196-55fe-462c-b012-be0d5fbe33ee-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.441883 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a38f1196-55fe-462c-b012-be0d5fbe33ee-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.441935 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9st4n\" (UniqueName: \"kubernetes.io/projected/a38f1196-55fe-462c-b012-be0d5fbe33ee-kube-api-access-9st4n\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.441970 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.442097 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a38f1196-55fe-462c-b012-be0d5fbe33ee-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.442130 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a38f1196-55fe-462c-b012-be0d5fbe33ee-config\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.442150 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a38f1196-55fe-462c-b012-be0d5fbe33ee-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.442185 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a38f1196-55fe-462c-b012-be0d5fbe33ee-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.443723 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.443929 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a38f1196-55fe-462c-b012-be0d5fbe33ee-config\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.445231 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a38f1196-55fe-462c-b012-be0d5fbe33ee-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.446214 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a38f1196-55fe-462c-b012-be0d5fbe33ee-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.451105 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a38f1196-55fe-462c-b012-be0d5fbe33ee-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.461303 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a38f1196-55fe-462c-b012-be0d5fbe33ee-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.483224 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a38f1196-55fe-462c-b012-be0d5fbe33ee-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.494760 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9st4n\" (UniqueName: \"kubernetes.io/projected/a38f1196-55fe-462c-b012-be0d5fbe33ee-kube-api-access-9st4n\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.505344 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a38f1196-55fe-462c-b012-be0d5fbe33ee\") " pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.522077 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543494 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-var-log-ovn\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543579 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fef2319a-5415-4712-a7c6-ad0ab4decff0-var-run\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543608 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-scripts\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543634 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fef2319a-5415-4712-a7c6-ad0ab4decff0-etc-ovs\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543660 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcn56\" (UniqueName: \"kubernetes.io/projected/fef2319a-5415-4712-a7c6-ad0ab4decff0-kube-api-access-xcn56\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543684 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fef2319a-5415-4712-a7c6-ad0ab4decff0-var-log\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543727 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whnj7\" (UniqueName: \"kubernetes.io/projected/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-kube-api-access-whnj7\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543754 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-var-run-ovn\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543779 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-combined-ca-bundle\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543798 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-ovn-controller-tls-certs\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543826 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fef2319a-5415-4712-a7c6-ad0ab4decff0-scripts\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543859 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fef2319a-5415-4712-a7c6-ad0ab4decff0-var-lib\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.543896 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-var-run\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.645757 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fef2319a-5415-4712-a7c6-ad0ab4decff0-var-run\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.645809 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-scripts\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.645837 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fef2319a-5415-4712-a7c6-ad0ab4decff0-etc-ovs\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.645859 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcn56\" (UniqueName: \"kubernetes.io/projected/fef2319a-5415-4712-a7c6-ad0ab4decff0-kube-api-access-xcn56\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.645986 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fef2319a-5415-4712-a7c6-ad0ab4decff0-var-log\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.646130 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whnj7\" (UniqueName: \"kubernetes.io/projected/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-kube-api-access-whnj7\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.646200 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-var-run-ovn\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.646231 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-combined-ca-bundle\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.646254 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-ovn-controller-tls-certs\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.646287 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fef2319a-5415-4712-a7c6-ad0ab4decff0-scripts\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.646334 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fef2319a-5415-4712-a7c6-ad0ab4decff0-var-lib\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.646401 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-var-run\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.646473 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-var-log-ovn\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.647083 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-var-log-ovn\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.647440 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-var-run-ovn\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.647570 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/fef2319a-5415-4712-a7c6-ad0ab4decff0-var-lib\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.647592 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/fef2319a-5415-4712-a7c6-ad0ab4decff0-var-log\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.647650 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-var-run\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.647779 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/fef2319a-5415-4712-a7c6-ad0ab4decff0-etc-ovs\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.648170 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fef2319a-5415-4712-a7c6-ad0ab4decff0-var-run\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.651579 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-scripts\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.652721 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fef2319a-5415-4712-a7c6-ad0ab4decff0-scripts\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.661314 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-combined-ca-bundle\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.664711 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-ovn-controller-tls-certs\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.669928 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcn56\" (UniqueName: \"kubernetes.io/projected/fef2319a-5415-4712-a7c6-ad0ab4decff0-kube-api-access-xcn56\") pod \"ovn-controller-ovs-ff8dm\" (UID: \"fef2319a-5415-4712-a7c6-ad0ab4decff0\") " pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.683845 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whnj7\" (UniqueName: \"kubernetes.io/projected/ad8606dc-c6d0-4ae7-905e-929d0a82ae8d-kube-api-access-whnj7\") pod \"ovn-controller-mhqr7\" (UID: \"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d\") " pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.743263 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:26:34 crc kubenswrapper[4744]: I1008 09:26:34.752665 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mhqr7" Oct 08 09:26:35 crc kubenswrapper[4744]: I1008 09:26:35.922390 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.624798 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.639583 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.639777 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.643766 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.645549 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vfz45" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.646871 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.647131 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.738875 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/815ecda8-d0f0-47c1-b638-11a5ca9fce02-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.738927 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ztgn\" (UniqueName: \"kubernetes.io/projected/815ecda8-d0f0-47c1-b638-11a5ca9fce02-kube-api-access-6ztgn\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.738959 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/815ecda8-d0f0-47c1-b638-11a5ca9fce02-config\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.738983 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815ecda8-d0f0-47c1-b638-11a5ca9fce02-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.739018 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/815ecda8-d0f0-47c1-b638-11a5ca9fce02-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.739039 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.739180 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/815ecda8-d0f0-47c1-b638-11a5ca9fce02-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.740283 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/815ecda8-d0f0-47c1-b638-11a5ca9fce02-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.842652 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ztgn\" (UniqueName: \"kubernetes.io/projected/815ecda8-d0f0-47c1-b638-11a5ca9fce02-kube-api-access-6ztgn\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.842714 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/815ecda8-d0f0-47c1-b638-11a5ca9fce02-config\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.842741 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815ecda8-d0f0-47c1-b638-11a5ca9fce02-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.842772 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/815ecda8-d0f0-47c1-b638-11a5ca9fce02-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.842794 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.842814 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/815ecda8-d0f0-47c1-b638-11a5ca9fce02-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.842855 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/815ecda8-d0f0-47c1-b638-11a5ca9fce02-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.842961 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/815ecda8-d0f0-47c1-b638-11a5ca9fce02-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.843807 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.844324 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/815ecda8-d0f0-47c1-b638-11a5ca9fce02-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.844797 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/815ecda8-d0f0-47c1-b638-11a5ca9fce02-config\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.845342 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/815ecda8-d0f0-47c1-b638-11a5ca9fce02-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.857255 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/815ecda8-d0f0-47c1-b638-11a5ca9fce02-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.857861 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/815ecda8-d0f0-47c1-b638-11a5ca9fce02-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.864625 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815ecda8-d0f0-47c1-b638-11a5ca9fce02-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.880946 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ztgn\" (UniqueName: \"kubernetes.io/projected/815ecda8-d0f0-47c1-b638-11a5ca9fce02-kube-api-access-6ztgn\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.898832 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"815ecda8-d0f0-47c1-b638-11a5ca9fce02\") " pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:37 crc kubenswrapper[4744]: I1008 09:26:37.980247 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 09:26:44 crc kubenswrapper[4744]: I1008 09:26:44.678405 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a38f1196-55fe-462c-b012-be0d5fbe33ee","Type":"ContainerStarted","Data":"7f633e9d51031d5df7fc563d830257ee9b1d0326101fe88d1b20b4ea1b45ff51"} Oct 08 09:26:45 crc kubenswrapper[4744]: I1008 09:26:45.405916 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mhqr7"] Oct 08 09:26:48 crc kubenswrapper[4744]: E1008 09:26:48.701740 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached@sha256:773daada6402d9cad089cdc809d6c0335456d057ac1a25441ab5d82add2f70f4" Oct 08 09:26:48 crc kubenswrapper[4744]: E1008 09:26:48.703074 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached@sha256:773daada6402d9cad089cdc809d6c0335456d057ac1a25441ab5d82add2f70f4,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n647h56ch67ch8fh5fbh55ch97h5b7hc4hdch587h58h5d7h574h578hd7h5fch577h649h684h659h5dbh5f9h668h649h5ffh5dh77h664h595h596h57bq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8b75c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(4c8e37bd-80bc-4e95-8baf-f571adecb232): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:26:48 crc kubenswrapper[4744]: E1008 09:26:48.704382 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="4c8e37bd-80bc-4e95-8baf-f571adecb232" Oct 08 09:26:48 crc kubenswrapper[4744]: E1008 09:26:48.720573 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached@sha256:773daada6402d9cad089cdc809d6c0335456d057ac1a25441ab5d82add2f70f4\\\"\"" pod="openstack/memcached-0" podUID="4c8e37bd-80bc-4e95-8baf-f571adecb232" Oct 08 09:26:49 crc kubenswrapper[4744]: I1008 09:26:49.690539 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:26:49 crc kubenswrapper[4744]: I1008 09:26:49.690595 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:26:49 crc kubenswrapper[4744]: E1008 09:26:49.860770 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0" Oct 08 09:26:49 crc kubenswrapper[4744]: E1008 09:26:49.860991 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lbscx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(6a6f5a1b-6abb-4568-89b5-a255f919cebb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:26:49 crc kubenswrapper[4744]: E1008 09:26:49.862191 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="6a6f5a1b-6abb-4568-89b5-a255f919cebb" Oct 08 09:26:50 crc kubenswrapper[4744]: E1008 09:26:50.737066 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0\\\"\"" pod="openstack/rabbitmq-server-0" podUID="6a6f5a1b-6abb-4568-89b5-a255f919cebb" Oct 08 09:26:51 crc kubenswrapper[4744]: W1008 09:26:51.666628 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad8606dc_c6d0_4ae7_905e_929d0a82ae8d.slice/crio-ef1ced30b16fa13e548108db1a8f6f0077fb93d41352f5147d5d32d8ecdb29c4 WatchSource:0}: Error finding container ef1ced30b16fa13e548108db1a8f6f0077fb93d41352f5147d5d32d8ecdb29c4: Status 404 returned error can't find the container with id ef1ced30b16fa13e548108db1a8f6f0077fb93d41352f5147d5d32d8ecdb29c4 Oct 08 09:26:51 crc kubenswrapper[4744]: E1008 09:26:51.706422 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0" Oct 08 09:26:51 crc kubenswrapper[4744]: E1008 09:26:51.706657 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dql57,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(18a53e2c-fcfa-4f49-aef0-ef1cee4118f6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:26:51 crc kubenswrapper[4744]: E1008 09:26:51.708122 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" Oct 08 09:26:51 crc kubenswrapper[4744]: I1008 09:26:51.747596 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mhqr7" event={"ID":"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d","Type":"ContainerStarted","Data":"ef1ced30b16fa13e548108db1a8f6f0077fb93d41352f5147d5d32d8ecdb29c4"} Oct 08 09:26:51 crc kubenswrapper[4744]: E1008 09:26:51.749594 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.580587 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.581572 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-shb5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-77597f887-ktnfw_openstack(00e8ebd9-71ce-42b9-8bd3-148d3f9309b9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.582869 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-77597f887-ktnfw" podUID="00e8ebd9-71ce-42b9-8bd3-148d3f9309b9" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.632670 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.632861 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l5ktb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-644597f84c-tcwbc_openstack(683a64cc-3b26-482b-a1bd-9eac5f056a4b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.634906 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-644597f84c-tcwbc" podUID="683a64cc-3b26-482b-a1bd-9eac5f056a4b" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.635871 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.636150 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v6f9w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bfcb9d745-9zj26_openstack(eca76c78-de7e-4eb0-8059-5a1fbcd99796): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.637331 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" podUID="eca76c78-de7e-4eb0-8059-5a1fbcd99796" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.688511 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.688732 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tqw9d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-758b79db4c-82cxq_openstack(4b19e1cb-2043-4cb1-b40f-2dc74cb6a011): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.690169 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-758b79db4c-82cxq" podUID="4b19e1cb-2043-4cb1-b40f-2dc74cb6a011" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.793929 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df\\\"\"" pod="openstack/dnsmasq-dns-77597f887-ktnfw" podUID="00e8ebd9-71ce-42b9-8bd3-148d3f9309b9" Oct 08 09:26:56 crc kubenswrapper[4744]: E1008 09:26:56.795096 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df\\\"\"" pod="openstack/dnsmasq-dns-644597f84c-tcwbc" podUID="683a64cc-3b26-482b-a1bd-9eac5f056a4b" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.280916 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.376652 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-ff8dm"] Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.646062 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-kshz8"] Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.654282 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.659762 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.677205 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/82e01314-590d-4fe4-9bbb-e5d180ff7093-ovs-rundir\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.677290 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/82e01314-590d-4fe4-9bbb-e5d180ff7093-ovn-rundir\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.677338 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/82e01314-590d-4fe4-9bbb-e5d180ff7093-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.677385 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e01314-590d-4fe4-9bbb-e5d180ff7093-combined-ca-bundle\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.677439 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zwz6\" (UniqueName: \"kubernetes.io/projected/82e01314-590d-4fe4-9bbb-e5d180ff7093-kube-api-access-4zwz6\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.677505 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82e01314-590d-4fe4-9bbb-e5d180ff7093-config\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.681837 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kshz8"] Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.779241 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82e01314-590d-4fe4-9bbb-e5d180ff7093-config\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.779620 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/82e01314-590d-4fe4-9bbb-e5d180ff7093-ovs-rundir\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.779682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/82e01314-590d-4fe4-9bbb-e5d180ff7093-ovn-rundir\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.779727 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/82e01314-590d-4fe4-9bbb-e5d180ff7093-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.779749 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e01314-590d-4fe4-9bbb-e5d180ff7093-combined-ca-bundle\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.779792 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zwz6\" (UniqueName: \"kubernetes.io/projected/82e01314-590d-4fe4-9bbb-e5d180ff7093-kube-api-access-4zwz6\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.780270 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82e01314-590d-4fe4-9bbb-e5d180ff7093-config\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.780607 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/82e01314-590d-4fe4-9bbb-e5d180ff7093-ovs-rundir\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.781132 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/82e01314-590d-4fe4-9bbb-e5d180ff7093-ovn-rundir\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.787626 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/82e01314-590d-4fe4-9bbb-e5d180ff7093-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.787766 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e01314-590d-4fe4-9bbb-e5d180ff7093-combined-ca-bundle\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.803803 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zwz6\" (UniqueName: \"kubernetes.io/projected/82e01314-590d-4fe4-9bbb-e5d180ff7093-kube-api-access-4zwz6\") pod \"ovn-controller-metrics-kshz8\" (UID: \"82e01314-590d-4fe4-9bbb-e5d180ff7093\") " pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.843469 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-tcwbc"] Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.923447 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-z5nmg"] Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.924965 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.926992 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-z5nmg"] Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.948772 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.984352 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh57v\" (UniqueName: \"kubernetes.io/projected/6471b061-3dd3-46b3-ac86-9051b15355fa-kube-api-access-sh57v\") pod \"dnsmasq-dns-6d8fd57975-z5nmg\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.984456 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-dns-svc\") pod \"dnsmasq-dns-6d8fd57975-z5nmg\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.984512 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8fd57975-z5nmg\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:57 crc kubenswrapper[4744]: I1008 09:26:57.984576 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-config\") pod \"dnsmasq-dns-6d8fd57975-z5nmg\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.001335 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kshz8" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.086782 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-config\") pod \"dnsmasq-dns-6d8fd57975-z5nmg\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.086859 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh57v\" (UniqueName: \"kubernetes.io/projected/6471b061-3dd3-46b3-ac86-9051b15355fa-kube-api-access-sh57v\") pod \"dnsmasq-dns-6d8fd57975-z5nmg\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.086922 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-dns-svc\") pod \"dnsmasq-dns-6d8fd57975-z5nmg\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.087002 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8fd57975-z5nmg\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.087843 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8fd57975-z5nmg\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.088453 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-dns-svc\") pod \"dnsmasq-dns-6d8fd57975-z5nmg\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.088549 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-config\") pod \"dnsmasq-dns-6d8fd57975-z5nmg\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.110436 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh57v\" (UniqueName: \"kubernetes.io/projected/6471b061-3dd3-46b3-ac86-9051b15355fa-kube-api-access-sh57v\") pod \"dnsmasq-dns-6d8fd57975-z5nmg\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.244130 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-ktnfw"] Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.279910 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.311247 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-lvxbs"] Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.314595 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.318413 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.327397 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-lvxbs"] Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.391991 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-ovsdbserver-nb\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.392066 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlwl2\" (UniqueName: \"kubernetes.io/projected/45d914c8-25de-4c20-84db-2d26c95289ee-kube-api-access-hlwl2\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.392098 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-ovsdbserver-sb\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.392144 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-dns-svc\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.392182 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-config\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.494412 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-config\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.496218 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-ovsdbserver-nb\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.496249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlwl2\" (UniqueName: \"kubernetes.io/projected/45d914c8-25de-4c20-84db-2d26c95289ee-kube-api-access-hlwl2\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.496284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-ovsdbserver-sb\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.496329 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-dns-svc\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.495931 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-config\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.498905 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-ovsdbserver-nb\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.498914 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-ovsdbserver-sb\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.499130 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-dns-svc\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.528432 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlwl2\" (UniqueName: \"kubernetes.io/projected/45d914c8-25de-4c20-84db-2d26c95289ee-kube-api-access-hlwl2\") pod \"dnsmasq-dns-bc45f6dcf-lvxbs\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: E1008 09:26:58.580907 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Oct 08 09:26:58 crc kubenswrapper[4744]: E1008 09:26:58.581056 4744 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb" Oct 08 09:26:58 crc kubenswrapper[4744]: E1008 09:26:58.581420 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pcm2r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(7eb871ad-c10b-48ea-b27e-f9dfca634845): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 09:26:58 crc kubenswrapper[4744]: E1008 09:26:58.582955 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="7eb871ad-c10b-48ea-b27e-f9dfca634845" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.653924 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.721468 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.737894 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.738539 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.741864 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.801404 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/683a64cc-3b26-482b-a1bd-9eac5f056a4b-dns-svc\") pod \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\" (UID: \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\") " Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.801451 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shb5b\" (UniqueName: \"kubernetes.io/projected/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-kube-api-access-shb5b\") pod \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\" (UID: \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\") " Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.801573 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-config\") pod \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\" (UID: \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\") " Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.801637 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6f9w\" (UniqueName: \"kubernetes.io/projected/eca76c78-de7e-4eb0-8059-5a1fbcd99796-kube-api-access-v6f9w\") pod \"eca76c78-de7e-4eb0-8059-5a1fbcd99796\" (UID: \"eca76c78-de7e-4eb0-8059-5a1fbcd99796\") " Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.801804 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-config\") pod \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\" (UID: \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\") " Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.801867 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca76c78-de7e-4eb0-8059-5a1fbcd99796-config\") pod \"eca76c78-de7e-4eb0-8059-5a1fbcd99796\" (UID: \"eca76c78-de7e-4eb0-8059-5a1fbcd99796\") " Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.801916 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-dns-svc\") pod \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\" (UID: \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\") " Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.801955 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-dns-svc\") pod \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\" (UID: \"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9\") " Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.801986 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/683a64cc-3b26-482b-a1bd-9eac5f056a4b-config\") pod \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\" (UID: \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\") " Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.802069 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5ktb\" (UniqueName: \"kubernetes.io/projected/683a64cc-3b26-482b-a1bd-9eac5f056a4b-kube-api-access-l5ktb\") pod \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\" (UID: \"683a64cc-3b26-482b-a1bd-9eac5f056a4b\") " Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.802123 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqw9d\" (UniqueName: \"kubernetes.io/projected/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-kube-api-access-tqw9d\") pod \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\" (UID: \"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011\") " Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.803993 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "00e8ebd9-71ce-42b9-8bd3-148d3f9309b9" (UID: "00e8ebd9-71ce-42b9-8bd3-148d3f9309b9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.804426 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/683a64cc-3b26-482b-a1bd-9eac5f056a4b-config" (OuterVolumeSpecName: "config") pod "683a64cc-3b26-482b-a1bd-9eac5f056a4b" (UID: "683a64cc-3b26-482b-a1bd-9eac5f056a4b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.804438 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eca76c78-de7e-4eb0-8059-5a1fbcd99796-config" (OuterVolumeSpecName: "config") pod "eca76c78-de7e-4eb0-8059-5a1fbcd99796" (UID: "eca76c78-de7e-4eb0-8059-5a1fbcd99796"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.804771 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4b19e1cb-2043-4cb1-b40f-2dc74cb6a011" (UID: "4b19e1cb-2043-4cb1-b40f-2dc74cb6a011"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.805247 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-config" (OuterVolumeSpecName: "config") pod "00e8ebd9-71ce-42b9-8bd3-148d3f9309b9" (UID: "00e8ebd9-71ce-42b9-8bd3-148d3f9309b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.805256 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/683a64cc-3b26-482b-a1bd-9eac5f056a4b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "683a64cc-3b26-482b-a1bd-9eac5f056a4b" (UID: "683a64cc-3b26-482b-a1bd-9eac5f056a4b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.806335 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-config" (OuterVolumeSpecName: "config") pod "4b19e1cb-2043-4cb1-b40f-2dc74cb6a011" (UID: "4b19e1cb-2043-4cb1-b40f-2dc74cb6a011"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.821250 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-kube-api-access-shb5b" (OuterVolumeSpecName: "kube-api-access-shb5b") pod "00e8ebd9-71ce-42b9-8bd3-148d3f9309b9" (UID: "00e8ebd9-71ce-42b9-8bd3-148d3f9309b9"). InnerVolumeSpecName "kube-api-access-shb5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.821307 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/683a64cc-3b26-482b-a1bd-9eac5f056a4b-kube-api-access-l5ktb" (OuterVolumeSpecName: "kube-api-access-l5ktb") pod "683a64cc-3b26-482b-a1bd-9eac5f056a4b" (UID: "683a64cc-3b26-482b-a1bd-9eac5f056a4b"). InnerVolumeSpecName "kube-api-access-l5ktb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.832756 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-kube-api-access-tqw9d" (OuterVolumeSpecName: "kube-api-access-tqw9d") pod "4b19e1cb-2043-4cb1-b40f-2dc74cb6a011" (UID: "4b19e1cb-2043-4cb1-b40f-2dc74cb6a011"). InnerVolumeSpecName "kube-api-access-tqw9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.846428 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eca76c78-de7e-4eb0-8059-5a1fbcd99796-kube-api-access-v6f9w" (OuterVolumeSpecName: "kube-api-access-v6f9w") pod "eca76c78-de7e-4eb0-8059-5a1fbcd99796" (UID: "eca76c78-de7e-4eb0-8059-5a1fbcd99796"). InnerVolumeSpecName "kube-api-access-v6f9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.886067 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-82cxq" event={"ID":"4b19e1cb-2043-4cb1-b40f-2dc74cb6a011","Type":"ContainerDied","Data":"52e7ada6a64094c67db720dda8c6e21d1a26de7353fee919ca06c40e15f1f90c"} Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.886226 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-82cxq" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.902656 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"815ecda8-d0f0-47c1-b638-11a5ca9fce02","Type":"ContainerStarted","Data":"541cb0608bcbabf4f275aaf7d7c18484e34dcb7e3feea87390fc6bb447c108e7"} Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.907472 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-ktnfw" event={"ID":"00e8ebd9-71ce-42b9-8bd3-148d3f9309b9","Type":"ContainerDied","Data":"0d321790518be440628efc9dae8c456dc785962d013bd221a6e93963d318f4dd"} Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.907793 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-ktnfw" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.910577 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-tcwbc" event={"ID":"683a64cc-3b26-482b-a1bd-9eac5f056a4b","Type":"ContainerDied","Data":"6f565aaf3ce43663611692c6974a8667b8c0613d4a252797c2b93c29575c61bf"} Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.913079 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-tcwbc" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.913444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" event={"ID":"eca76c78-de7e-4eb0-8059-5a1fbcd99796","Type":"ContainerDied","Data":"0a1cc0bb614874c8d23e9493b3c2497f68e4f69243365ff6863ed311813231d3"} Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.914490 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-9zj26" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.914966 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.920342 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ff8dm" event={"ID":"fef2319a-5415-4712-a7c6-ad0ab4decff0","Type":"ContainerStarted","Data":"debb3407b85f705511ef6ca08f573c61456aff40fee8cb626fc91500e8cdf4f3"} Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.934123 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca76c78-de7e-4eb0-8059-5a1fbcd99796-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.950774 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.950938 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.951033 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/683a64cc-3b26-482b-a1bd-9eac5f056a4b-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.951138 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5ktb\" (UniqueName: \"kubernetes.io/projected/683a64cc-3b26-482b-a1bd-9eac5f056a4b-kube-api-access-l5ktb\") on node \"crc\" DevicePath \"\"" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.951206 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqw9d\" (UniqueName: \"kubernetes.io/projected/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011-kube-api-access-tqw9d\") on node \"crc\" DevicePath \"\"" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.951287 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/683a64cc-3b26-482b-a1bd-9eac5f056a4b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.951416 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shb5b\" (UniqueName: \"kubernetes.io/projected/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-kube-api-access-shb5b\") on node \"crc\" DevicePath \"\"" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.951486 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:26:58 crc kubenswrapper[4744]: I1008 09:26:58.951545 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6f9w\" (UniqueName: \"kubernetes.io/projected/eca76c78-de7e-4eb0-8059-5a1fbcd99796-kube-api-access-v6f9w\") on node \"crc\" DevicePath \"\"" Oct 08 09:26:58 crc kubenswrapper[4744]: E1008 09:26:58.969382 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb\\\"\"" pod="openstack/kube-state-metrics-0" podUID="7eb871ad-c10b-48ea-b27e-f9dfca634845" Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.028222 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kshz8"] Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.071117 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-tcwbc"] Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.089268 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-tcwbc"] Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.167579 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-9zj26"] Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.193320 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-9zj26"] Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.221849 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-82cxq"] Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.228485 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-82cxq"] Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.243877 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-ktnfw"] Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.250047 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77597f887-ktnfw"] Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.269291 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-lvxbs"] Oct 08 09:26:59 crc kubenswrapper[4744]: W1008 09:26:59.276739 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45d914c8_25de_4c20_84db_2d26c95289ee.slice/crio-e3790c372ceb3218d3b98287b21e9a25458e5f97d92deab71ec6b94e505a0aca WatchSource:0}: Error finding container e3790c372ceb3218d3b98287b21e9a25458e5f97d92deab71ec6b94e505a0aca: Status 404 returned error can't find the container with id e3790c372ceb3218d3b98287b21e9a25458e5f97d92deab71ec6b94e505a0aca Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.320822 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-z5nmg"] Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.503424 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00e8ebd9-71ce-42b9-8bd3-148d3f9309b9" path="/var/lib/kubelet/pods/00e8ebd9-71ce-42b9-8bd3-148d3f9309b9/volumes" Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.504099 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b19e1cb-2043-4cb1-b40f-2dc74cb6a011" path="/var/lib/kubelet/pods/4b19e1cb-2043-4cb1-b40f-2dc74cb6a011/volumes" Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.504567 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="683a64cc-3b26-482b-a1bd-9eac5f056a4b" path="/var/lib/kubelet/pods/683a64cc-3b26-482b-a1bd-9eac5f056a4b/volumes" Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.505019 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eca76c78-de7e-4eb0-8059-5a1fbcd99796" path="/var/lib/kubelet/pods/eca76c78-de7e-4eb0-8059-5a1fbcd99796/volumes" Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.932865 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" event={"ID":"6471b061-3dd3-46b3-ac86-9051b15355fa","Type":"ContainerStarted","Data":"8c0acf8f9e452f2aff13aba0579aca8b5e810abce2f4d513c1d7f4760ddf5690"} Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.935850 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d1412b08-0e26-455c-afb5-3f51b2cb3012","Type":"ContainerStarted","Data":"0109442e912f4c8d711f468d824fd1c8e503e7335ed7be2fc0fefcdf359d74ce"} Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.943537 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ef3dc460-b269-4e4c-b1db-bfb550b58a7b","Type":"ContainerStarted","Data":"03a805d8ed330f1eb1c52cda4f560c1d74edd64cb67c31ea7f39e2c8da5c136f"} Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.945671 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" event={"ID":"45d914c8-25de-4c20-84db-2d26c95289ee","Type":"ContainerStarted","Data":"e3790c372ceb3218d3b98287b21e9a25458e5f97d92deab71ec6b94e505a0aca"} Oct 08 09:26:59 crc kubenswrapper[4744]: I1008 09:26:59.948095 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kshz8" event={"ID":"82e01314-590d-4fe4-9bbb-e5d180ff7093","Type":"ContainerStarted","Data":"3a58443a54598aef7f5978ac77dcb2bf1a30f682f5bdcd2a455a4bd2b43e03a4"} Oct 08 09:27:07 crc kubenswrapper[4744]: I1008 09:27:07.006793 4744 generic.go:334] "Generic (PLEG): container finished" podID="d1412b08-0e26-455c-afb5-3f51b2cb3012" containerID="0109442e912f4c8d711f468d824fd1c8e503e7335ed7be2fc0fefcdf359d74ce" exitCode=0 Oct 08 09:27:07 crc kubenswrapper[4744]: I1008 09:27:07.006851 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d1412b08-0e26-455c-afb5-3f51b2cb3012","Type":"ContainerDied","Data":"0109442e912f4c8d711f468d824fd1c8e503e7335ed7be2fc0fefcdf359d74ce"} Oct 08 09:27:07 crc kubenswrapper[4744]: I1008 09:27:07.010496 4744 generic.go:334] "Generic (PLEG): container finished" podID="ef3dc460-b269-4e4c-b1db-bfb550b58a7b" containerID="03a805d8ed330f1eb1c52cda4f560c1d74edd64cb67c31ea7f39e2c8da5c136f" exitCode=0 Oct 08 09:27:07 crc kubenswrapper[4744]: I1008 09:27:07.010555 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ef3dc460-b269-4e4c-b1db-bfb550b58a7b","Type":"ContainerDied","Data":"03a805d8ed330f1eb1c52cda4f560c1d74edd64cb67c31ea7f39e2c8da5c136f"} Oct 08 09:27:08 crc kubenswrapper[4744]: I1008 09:27:08.020332 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ff8dm" event={"ID":"fef2319a-5415-4712-a7c6-ad0ab4decff0","Type":"ContainerStarted","Data":"7b5c537ed20c5e2e39a99f113dbb82914df09935a1bad20f739e6000c239eba6"} Oct 08 09:27:08 crc kubenswrapper[4744]: I1008 09:27:08.024740 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d1412b08-0e26-455c-afb5-3f51b2cb3012","Type":"ContainerStarted","Data":"189c2cd3c502b6b61956e3a5586d713e92f7f9b069bd71580b6429e73812c186"} Oct 08 09:27:08 crc kubenswrapper[4744]: I1008 09:27:08.027411 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ef3dc460-b269-4e4c-b1db-bfb550b58a7b","Type":"ContainerStarted","Data":"161d3b31041bdf6e3b2afc0b955c563c20f51a4ef900a0715aae98639e6ff5ad"} Oct 08 09:27:08 crc kubenswrapper[4744]: I1008 09:27:08.061902 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=13.521826322999999 podStartE2EDuration="42.061882148s" podCreationTimestamp="2025-10-08 09:26:26 +0000 UTC" firstStartedPulling="2025-10-08 09:26:29.988546994 +0000 UTC m=+885.236192233" lastFinishedPulling="2025-10-08 09:26:58.528602819 +0000 UTC m=+913.776248058" observedRunningTime="2025-10-08 09:27:08.058269945 +0000 UTC m=+923.305915214" watchObservedRunningTime="2025-10-08 09:27:08.061882148 +0000 UTC m=+923.309527387" Oct 08 09:27:08 crc kubenswrapper[4744]: I1008 09:27:08.078994 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=14.000824251 podStartE2EDuration="41.078978285s" podCreationTimestamp="2025-10-08 09:26:27 +0000 UTC" firstStartedPulling="2025-10-08 09:26:30.038553313 +0000 UTC m=+885.286198552" lastFinishedPulling="2025-10-08 09:26:57.116707347 +0000 UTC m=+912.364352586" observedRunningTime="2025-10-08 09:27:08.077926666 +0000 UTC m=+923.325571915" watchObservedRunningTime="2025-10-08 09:27:08.078978285 +0000 UTC m=+923.326623524" Oct 08 09:27:08 crc kubenswrapper[4744]: I1008 09:27:08.677765 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 08 09:27:08 crc kubenswrapper[4744]: I1008 09:27:08.678113 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 08 09:27:08 crc kubenswrapper[4744]: I1008 09:27:08.863486 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 08 09:27:08 crc kubenswrapper[4744]: I1008 09:27:08.863541 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.036921 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4c8e37bd-80bc-4e95-8baf-f571adecb232","Type":"ContainerStarted","Data":"c85d5d7ca900b28c8a0a6e765d682a6cc5b5ae358ebaf55bc17752819f0ba10b"} Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.037154 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.039413 4744 generic.go:334] "Generic (PLEG): container finished" podID="fef2319a-5415-4712-a7c6-ad0ab4decff0" containerID="7b5c537ed20c5e2e39a99f113dbb82914df09935a1bad20f739e6000c239eba6" exitCode=0 Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.039507 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ff8dm" event={"ID":"fef2319a-5415-4712-a7c6-ad0ab4decff0","Type":"ContainerDied","Data":"7b5c537ed20c5e2e39a99f113dbb82914df09935a1bad20f739e6000c239eba6"} Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.041832 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"815ecda8-d0f0-47c1-b638-11a5ca9fce02","Type":"ContainerStarted","Data":"102fcf2f12554e027ce7b6720dcb4f6cf1b018a21e04a5319787bbf25cde0f03"} Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.043293 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6","Type":"ContainerStarted","Data":"a7ae4524f517636f791ec2332601150d328e8d3eb0890dfd3c61265b8ed54e20"} Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.045189 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a38f1196-55fe-462c-b012-be0d5fbe33ee","Type":"ContainerStarted","Data":"dd27fa0d8a69e34b1b58b8c542bb5c0784fa4cd1388e809f88ca2bb9ef7ae7d1"} Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.047669 4744 generic.go:334] "Generic (PLEG): container finished" podID="6471b061-3dd3-46b3-ac86-9051b15355fa" containerID="7afee42ccc67c77d229b6d74efe4c006d82b31a7f036f401fd258ff844240111" exitCode=0 Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.047747 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" event={"ID":"6471b061-3dd3-46b3-ac86-9051b15355fa","Type":"ContainerDied","Data":"7afee42ccc67c77d229b6d74efe4c006d82b31a7f036f401fd258ff844240111"} Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.049427 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mhqr7" event={"ID":"ad8606dc-c6d0-4ae7-905e-929d0a82ae8d","Type":"ContainerStarted","Data":"77ea65449c0a02357cec8b9f0936de7139b3d2aed4c135aa60ed1e3217d415ce"} Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.049609 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-mhqr7" Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.051194 4744 generic.go:334] "Generic (PLEG): container finished" podID="45d914c8-25de-4c20-84db-2d26c95289ee" containerID="89ef0bc8c22be55fcfb810002c146fc9173bbec232942305725d3d8739ef3622" exitCode=0 Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.051235 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" event={"ID":"45d914c8-25de-4c20-84db-2d26c95289ee","Type":"ContainerDied","Data":"89ef0bc8c22be55fcfb810002c146fc9173bbec232942305725d3d8739ef3622"} Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.060142 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.62232543 podStartE2EDuration="41.060120807s" podCreationTimestamp="2025-10-08 09:26:28 +0000 UTC" firstStartedPulling="2025-10-08 09:26:30.09160569 +0000 UTC m=+885.339250929" lastFinishedPulling="2025-10-08 09:27:07.529401067 +0000 UTC m=+922.777046306" observedRunningTime="2025-10-08 09:27:09.0581066 +0000 UTC m=+924.305751849" watchObservedRunningTime="2025-10-08 09:27:09.060120807 +0000 UTC m=+924.307766046" Oct 08 09:27:09 crc kubenswrapper[4744]: I1008 09:27:09.130451 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-mhqr7" podStartSLOduration=19.484892387 podStartE2EDuration="35.130430421s" podCreationTimestamp="2025-10-08 09:26:34 +0000 UTC" firstStartedPulling="2025-10-08 09:26:51.68381736 +0000 UTC m=+906.931462599" lastFinishedPulling="2025-10-08 09:27:07.329355394 +0000 UTC m=+922.577000633" observedRunningTime="2025-10-08 09:27:09.121131206 +0000 UTC m=+924.368776445" watchObservedRunningTime="2025-10-08 09:27:09.130430421 +0000 UTC m=+924.378075660" Oct 08 09:27:10 crc kubenswrapper[4744]: I1008 09:27:10.066209 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6a6f5a1b-6abb-4568-89b5-a255f919cebb","Type":"ContainerStarted","Data":"816a11709d28ae06d2f534850a7bff8d19e598d0723d36416765dd4b97d1863e"} Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.094102 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7eb871ad-c10b-48ea-b27e-f9dfca634845","Type":"ContainerStarted","Data":"54d39aa20d340d7dce09a9f85db69d33883b237537de14c0bde01b8fc28ad6f7"} Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.095027 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.097709 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a38f1196-55fe-462c-b012-be0d5fbe33ee","Type":"ContainerStarted","Data":"80d209cd7a5f8c231275464104fdc90187ed99733aa7dd9da7af45b6edb344f9"} Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.107692 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kshz8" event={"ID":"82e01314-590d-4fe4-9bbb-e5d180ff7093","Type":"ContainerStarted","Data":"f6d1227de5195c3324ce3c25f0e4f41a6bc46e98769666680071114054a5305d"} Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.113389 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.537170867 podStartE2EDuration="42.113363213s" podCreationTimestamp="2025-10-08 09:26:30 +0000 UTC" firstStartedPulling="2025-10-08 09:26:31.588694527 +0000 UTC m=+886.836339766" lastFinishedPulling="2025-10-08 09:27:11.164886873 +0000 UTC m=+926.412532112" observedRunningTime="2025-10-08 09:27:12.111134789 +0000 UTC m=+927.358780038" watchObservedRunningTime="2025-10-08 09:27:12.113363213 +0000 UTC m=+927.361008452" Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.115401 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" event={"ID":"6471b061-3dd3-46b3-ac86-9051b15355fa","Type":"ContainerStarted","Data":"2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c"} Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.115960 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.131072 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" event={"ID":"45d914c8-25de-4c20-84db-2d26c95289ee","Type":"ContainerStarted","Data":"9b2405648a5422cb7d36175a47c300fe650aeb02ee481e5cd9de79b6929756eb"} Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.131166 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.135658 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-kshz8" podStartSLOduration=3.029221011 podStartE2EDuration="15.135629908s" podCreationTimestamp="2025-10-08 09:26:57 +0000 UTC" firstStartedPulling="2025-10-08 09:26:59.102949794 +0000 UTC m=+914.350595033" lastFinishedPulling="2025-10-08 09:27:11.209358691 +0000 UTC m=+926.457003930" observedRunningTime="2025-10-08 09:27:12.131350746 +0000 UTC m=+927.378995995" watchObservedRunningTime="2025-10-08 09:27:12.135629908 +0000 UTC m=+927.383275147" Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.167776 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ff8dm" event={"ID":"fef2319a-5415-4712-a7c6-ad0ab4decff0","Type":"ContainerStarted","Data":"f47cba17eb3ba5fbd957d2957defd07495506206c8394c23df293e2a3f97c955"} Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.167828 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-ff8dm" event={"ID":"fef2319a-5415-4712-a7c6-ad0ab4decff0","Type":"ContainerStarted","Data":"72c31f25827fb0514ad715ea9040f3c17fd04a6ed90e4b9dd873d89ae873b214"} Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.168612 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.168649 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.174226 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"815ecda8-d0f0-47c1-b638-11a5ca9fce02","Type":"ContainerStarted","Data":"132db4a6dce9be43dac88f0fa5f6337a9d6390cec88d5bcc3c34fa267ec7adff"} Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.181533 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=11.966996071 podStartE2EDuration="39.181509567s" podCreationTimestamp="2025-10-08 09:26:33 +0000 UTC" firstStartedPulling="2025-10-08 09:26:43.829882722 +0000 UTC m=+899.077527971" lastFinishedPulling="2025-10-08 09:27:11.044396228 +0000 UTC m=+926.292041467" observedRunningTime="2025-10-08 09:27:12.181147645 +0000 UTC m=+927.428792894" watchObservedRunningTime="2025-10-08 09:27:12.181509567 +0000 UTC m=+927.429154806" Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.215912 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" podStartSLOduration=6.015902329 podStartE2EDuration="14.215883956s" podCreationTimestamp="2025-10-08 09:26:58 +0000 UTC" firstStartedPulling="2025-10-08 09:26:59.289578675 +0000 UTC m=+914.537223914" lastFinishedPulling="2025-10-08 09:27:07.489560302 +0000 UTC m=+922.737205541" observedRunningTime="2025-10-08 09:27:12.20970156 +0000 UTC m=+927.457346809" watchObservedRunningTime="2025-10-08 09:27:12.215883956 +0000 UTC m=+927.463529195" Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.236567 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=23.80471276 podStartE2EDuration="36.236548885s" podCreationTimestamp="2025-10-08 09:26:36 +0000 UTC" firstStartedPulling="2025-10-08 09:26:58.613730436 +0000 UTC m=+913.861375675" lastFinishedPulling="2025-10-08 09:27:11.045566561 +0000 UTC m=+926.293211800" observedRunningTime="2025-10-08 09:27:12.234571539 +0000 UTC m=+927.482216778" watchObservedRunningTime="2025-10-08 09:27:12.236548885 +0000 UTC m=+927.484194124" Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.270799 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-ff8dm" podStartSLOduration=29.634720021 podStartE2EDuration="38.270768781s" podCreationTimestamp="2025-10-08 09:26:34 +0000 UTC" firstStartedPulling="2025-10-08 09:26:58.67174915 +0000 UTC m=+913.919394389" lastFinishedPulling="2025-10-08 09:27:07.30779791 +0000 UTC m=+922.555443149" observedRunningTime="2025-10-08 09:27:12.261003983 +0000 UTC m=+927.508649232" watchObservedRunningTime="2025-10-08 09:27:12.270768781 +0000 UTC m=+927.518414020" Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.297622 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" podStartSLOduration=7.10116648 podStartE2EDuration="15.297602686s" podCreationTimestamp="2025-10-08 09:26:57 +0000 UTC" firstStartedPulling="2025-10-08 09:26:59.338392866 +0000 UTC m=+914.586038105" lastFinishedPulling="2025-10-08 09:27:07.534829072 +0000 UTC m=+922.782474311" observedRunningTime="2025-10-08 09:27:12.294609951 +0000 UTC m=+927.542255180" watchObservedRunningTime="2025-10-08 09:27:12.297602686 +0000 UTC m=+927.545247925" Oct 08 09:27:12 crc kubenswrapper[4744]: I1008 09:27:12.981424 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 08 09:27:13 crc kubenswrapper[4744]: I1008 09:27:13.013752 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 08 09:27:13 crc kubenswrapper[4744]: I1008 09:27:13.066408 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 08 09:27:13 crc kubenswrapper[4744]: I1008 09:27:13.523161 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 08 09:27:13 crc kubenswrapper[4744]: I1008 09:27:13.570180 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 08 09:27:13 crc kubenswrapper[4744]: I1008 09:27:13.981353 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.026720 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.099676 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.191270 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.236012 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.293464 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.431605 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-r4dmt"] Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.434147 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-r4dmt" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.457822 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-r4dmt"] Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.477185 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q64lz\" (UniqueName: \"kubernetes.io/projected/d700c659-73db-4dc3-a2de-37e754f2775e-kube-api-access-q64lz\") pod \"glance-db-create-r4dmt\" (UID: \"d700c659-73db-4dc3-a2de-37e754f2775e\") " pod="openstack/glance-db-create-r4dmt" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.578902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q64lz\" (UniqueName: \"kubernetes.io/projected/d700c659-73db-4dc3-a2de-37e754f2775e-kube-api-access-q64lz\") pod \"glance-db-create-r4dmt\" (UID: \"d700c659-73db-4dc3-a2de-37e754f2775e\") " pod="openstack/glance-db-create-r4dmt" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.609114 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q64lz\" (UniqueName: \"kubernetes.io/projected/d700c659-73db-4dc3-a2de-37e754f2775e-kube-api-access-q64lz\") pod \"glance-db-create-r4dmt\" (UID: \"d700c659-73db-4dc3-a2de-37e754f2775e\") " pod="openstack/glance-db-create-r4dmt" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.679311 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.680821 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.684847 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.685035 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-mr4xs" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.687610 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.687955 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.708550 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.759435 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-r4dmt" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.781585 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.781643 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.781661 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.781680 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcbrc\" (UniqueName: \"kubernetes.io/projected/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-kube-api-access-dcbrc\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.781735 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-scripts\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.781759 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-config\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.781817 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.883658 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-config\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.884106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.884171 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.884220 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.884242 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.884270 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcbrc\" (UniqueName: \"kubernetes.io/projected/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-kube-api-access-dcbrc\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.884333 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-scripts\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.884844 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-config\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.885088 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.885498 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-scripts\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.888887 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.890028 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.904860 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:14 crc kubenswrapper[4744]: I1008 09:27:14.907614 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcbrc\" (UniqueName: \"kubernetes.io/projected/2c0cba37-bef9-4b03-8c6a-43acd36d04ff-kube-api-access-dcbrc\") pod \"ovn-northd-0\" (UID: \"2c0cba37-bef9-4b03-8c6a-43acd36d04ff\") " pod="openstack/ovn-northd-0" Oct 08 09:27:15 crc kubenswrapper[4744]: I1008 09:27:15.005152 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 09:27:15 crc kubenswrapper[4744]: I1008 09:27:15.027047 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 08 09:27:15 crc kubenswrapper[4744]: I1008 09:27:15.119913 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 08 09:27:15 crc kubenswrapper[4744]: I1008 09:27:15.339109 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-r4dmt"] Oct 08 09:27:15 crc kubenswrapper[4744]: I1008 09:27:15.511678 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 09:27:15 crc kubenswrapper[4744]: W1008 09:27:15.515294 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c0cba37_bef9_4b03_8c6a_43acd36d04ff.slice/crio-f0f723783a8faeb24d6d665494a780a0653cdea6100c179d22c1fed02422a16a WatchSource:0}: Error finding container f0f723783a8faeb24d6d665494a780a0653cdea6100c179d22c1fed02422a16a: Status 404 returned error can't find the container with id f0f723783a8faeb24d6d665494a780a0653cdea6100c179d22c1fed02422a16a Oct 08 09:27:16 crc kubenswrapper[4744]: I1008 09:27:16.206949 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2c0cba37-bef9-4b03-8c6a-43acd36d04ff","Type":"ContainerStarted","Data":"f0f723783a8faeb24d6d665494a780a0653cdea6100c179d22c1fed02422a16a"} Oct 08 09:27:16 crc kubenswrapper[4744]: I1008 09:27:16.209242 4744 generic.go:334] "Generic (PLEG): container finished" podID="d700c659-73db-4dc3-a2de-37e754f2775e" containerID="977cd4c09fdbc0d482979d07803fa655b3235316582e09225a68e1e2e243e9cb" exitCode=0 Oct 08 09:27:16 crc kubenswrapper[4744]: I1008 09:27:16.210534 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-r4dmt" event={"ID":"d700c659-73db-4dc3-a2de-37e754f2775e","Type":"ContainerDied","Data":"977cd4c09fdbc0d482979d07803fa655b3235316582e09225a68e1e2e243e9cb"} Oct 08 09:27:16 crc kubenswrapper[4744]: I1008 09:27:16.210605 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-r4dmt" event={"ID":"d700c659-73db-4dc3-a2de-37e754f2775e","Type":"ContainerStarted","Data":"827e4c284b7f0126f7d86655e9185879a1709f7718dc3cf7be20d8eb2e790d86"} Oct 08 09:27:17 crc kubenswrapper[4744]: I1008 09:27:17.221283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2c0cba37-bef9-4b03-8c6a-43acd36d04ff","Type":"ContainerStarted","Data":"796f4c1798ab687f79871049c7d57889dc66c0589d2ef204a75d3024c3e4a061"} Oct 08 09:27:17 crc kubenswrapper[4744]: I1008 09:27:17.221716 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"2c0cba37-bef9-4b03-8c6a-43acd36d04ff","Type":"ContainerStarted","Data":"52f130c1a1e81bf2d990047a5abb5e0252dbbcbf951211077a37d5c14de7c55c"} Oct 08 09:27:17 crc kubenswrapper[4744]: I1008 09:27:17.253690 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.191526339 podStartE2EDuration="3.25366718s" podCreationTimestamp="2025-10-08 09:27:14 +0000 UTC" firstStartedPulling="2025-10-08 09:27:15.517893155 +0000 UTC m=+930.765538394" lastFinishedPulling="2025-10-08 09:27:16.580033996 +0000 UTC m=+931.827679235" observedRunningTime="2025-10-08 09:27:17.248707498 +0000 UTC m=+932.496352787" watchObservedRunningTime="2025-10-08 09:27:17.25366718 +0000 UTC m=+932.501312419" Oct 08 09:27:17 crc kubenswrapper[4744]: I1008 09:27:17.579849 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-r4dmt" Oct 08 09:27:17 crc kubenswrapper[4744]: I1008 09:27:17.735421 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q64lz\" (UniqueName: \"kubernetes.io/projected/d700c659-73db-4dc3-a2de-37e754f2775e-kube-api-access-q64lz\") pod \"d700c659-73db-4dc3-a2de-37e754f2775e\" (UID: \"d700c659-73db-4dc3-a2de-37e754f2775e\") " Oct 08 09:27:17 crc kubenswrapper[4744]: I1008 09:27:17.742593 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d700c659-73db-4dc3-a2de-37e754f2775e-kube-api-access-q64lz" (OuterVolumeSpecName: "kube-api-access-q64lz") pod "d700c659-73db-4dc3-a2de-37e754f2775e" (UID: "d700c659-73db-4dc3-a2de-37e754f2775e"). InnerVolumeSpecName "kube-api-access-q64lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:27:17 crc kubenswrapper[4744]: I1008 09:27:17.838154 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q64lz\" (UniqueName: \"kubernetes.io/projected/d700c659-73db-4dc3-a2de-37e754f2775e-kube-api-access-q64lz\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.231483 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-r4dmt" Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.231518 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-r4dmt" event={"ID":"d700c659-73db-4dc3-a2de-37e754f2775e","Type":"ContainerDied","Data":"827e4c284b7f0126f7d86655e9185879a1709f7718dc3cf7be20d8eb2e790d86"} Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.233176 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="827e4c284b7f0126f7d86655e9185879a1709f7718dc3cf7be20d8eb2e790d86" Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.233206 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.283833 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.656361 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.724152 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-z5nmg"] Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.799701 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-rdzjs"] Oct 08 09:27:18 crc kubenswrapper[4744]: E1008 09:27:18.800220 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d700c659-73db-4dc3-a2de-37e754f2775e" containerName="mariadb-database-create" Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.800252 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d700c659-73db-4dc3-a2de-37e754f2775e" containerName="mariadb-database-create" Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.800686 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d700c659-73db-4dc3-a2de-37e754f2775e" containerName="mariadb-database-create" Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.802589 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-rdzjs" Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.809629 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-rdzjs"] Oct 08 09:27:18 crc kubenswrapper[4744]: I1008 09:27:18.976547 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nklg6\" (UniqueName: \"kubernetes.io/projected/1d23fe19-d889-4aa8-9cc9-35eecb6449bc-kube-api-access-nklg6\") pod \"keystone-db-create-rdzjs\" (UID: \"1d23fe19-d889-4aa8-9cc9-35eecb6449bc\") " pod="openstack/keystone-db-create-rdzjs" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.079172 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nklg6\" (UniqueName: \"kubernetes.io/projected/1d23fe19-d889-4aa8-9cc9-35eecb6449bc-kube-api-access-nklg6\") pod \"keystone-db-create-rdzjs\" (UID: \"1d23fe19-d889-4aa8-9cc9-35eecb6449bc\") " pod="openstack/keystone-db-create-rdzjs" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.115037 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nklg6\" (UniqueName: \"kubernetes.io/projected/1d23fe19-d889-4aa8-9cc9-35eecb6449bc-kube-api-access-nklg6\") pod \"keystone-db-create-rdzjs\" (UID: \"1d23fe19-d889-4aa8-9cc9-35eecb6449bc\") " pod="openstack/keystone-db-create-rdzjs" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.123463 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-rdzjs" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.200949 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-t9jv2"] Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.201964 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-t9jv2" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.209865 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-t9jv2"] Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.243933 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" podUID="6471b061-3dd3-46b3-ac86-9051b15355fa" containerName="dnsmasq-dns" containerID="cri-o://2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c" gracePeriod=10 Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.390952 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6brcm\" (UniqueName: \"kubernetes.io/projected/744dc616-99e5-4b30-bd16-476de000e0ae-kube-api-access-6brcm\") pod \"placement-db-create-t9jv2\" (UID: \"744dc616-99e5-4b30-bd16-476de000e0ae\") " pod="openstack/placement-db-create-t9jv2" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.494324 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6brcm\" (UniqueName: \"kubernetes.io/projected/744dc616-99e5-4b30-bd16-476de000e0ae-kube-api-access-6brcm\") pod \"placement-db-create-t9jv2\" (UID: \"744dc616-99e5-4b30-bd16-476de000e0ae\") " pod="openstack/placement-db-create-t9jv2" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.514483 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6brcm\" (UniqueName: \"kubernetes.io/projected/744dc616-99e5-4b30-bd16-476de000e0ae-kube-api-access-6brcm\") pod \"placement-db-create-t9jv2\" (UID: \"744dc616-99e5-4b30-bd16-476de000e0ae\") " pod="openstack/placement-db-create-t9jv2" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.558832 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-t9jv2" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.647527 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-rdzjs"] Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.689829 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.689885 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.689936 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.690566 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06edcaf88cdb63c2778de4ebf46b1539980361ea9f30176a88a8565819ce43a4"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.690611 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://06edcaf88cdb63c2778de4ebf46b1539980361ea9f30176a88a8565819ce43a4" gracePeriod=600 Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.709972 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.906648 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-config\") pod \"6471b061-3dd3-46b3-ac86-9051b15355fa\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.906813 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-ovsdbserver-nb\") pod \"6471b061-3dd3-46b3-ac86-9051b15355fa\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.906924 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh57v\" (UniqueName: \"kubernetes.io/projected/6471b061-3dd3-46b3-ac86-9051b15355fa-kube-api-access-sh57v\") pod \"6471b061-3dd3-46b3-ac86-9051b15355fa\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.906952 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-dns-svc\") pod \"6471b061-3dd3-46b3-ac86-9051b15355fa\" (UID: \"6471b061-3dd3-46b3-ac86-9051b15355fa\") " Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.918764 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6471b061-3dd3-46b3-ac86-9051b15355fa-kube-api-access-sh57v" (OuterVolumeSpecName: "kube-api-access-sh57v") pod "6471b061-3dd3-46b3-ac86-9051b15355fa" (UID: "6471b061-3dd3-46b3-ac86-9051b15355fa"). InnerVolumeSpecName "kube-api-access-sh57v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.963828 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-config" (OuterVolumeSpecName: "config") pod "6471b061-3dd3-46b3-ac86-9051b15355fa" (UID: "6471b061-3dd3-46b3-ac86-9051b15355fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.968727 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6471b061-3dd3-46b3-ac86-9051b15355fa" (UID: "6471b061-3dd3-46b3-ac86-9051b15355fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:27:19 crc kubenswrapper[4744]: I1008 09:27:19.979625 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6471b061-3dd3-46b3-ac86-9051b15355fa" (UID: "6471b061-3dd3-46b3-ac86-9051b15355fa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.012115 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.012308 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.012405 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh57v\" (UniqueName: \"kubernetes.io/projected/6471b061-3dd3-46b3-ac86-9051b15355fa-kube-api-access-sh57v\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.012508 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6471b061-3dd3-46b3-ac86-9051b15355fa-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.107609 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-t9jv2"] Oct 08 09:27:20 crc kubenswrapper[4744]: E1008 09:27:20.125818 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d23fe19_d889_4aa8_9cc9_35eecb6449bc.slice/crio-e65f207bcf2f14ccb82b865966626dc284f7f15e57faebc3d79703a496df6b10.scope\": RecentStats: unable to find data in memory cache]" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.263314 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="06edcaf88cdb63c2778de4ebf46b1539980361ea9f30176a88a8565819ce43a4" exitCode=0 Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.263815 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"06edcaf88cdb63c2778de4ebf46b1539980361ea9f30176a88a8565819ce43a4"} Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.263844 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"fc794dea8745685aa845397cb02ef8c3c9ae63ea691d5f60cd28e1c653036510"} Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.263862 4744 scope.go:117] "RemoveContainer" containerID="12c8d8cd4abdd9a59c0cfbc9852b990e3820f44c3e436a88bb84a00c67314d34" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.269798 4744 generic.go:334] "Generic (PLEG): container finished" podID="6471b061-3dd3-46b3-ac86-9051b15355fa" containerID="2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c" exitCode=0 Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.269933 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" event={"ID":"6471b061-3dd3-46b3-ac86-9051b15355fa","Type":"ContainerDied","Data":"2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c"} Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.270008 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" event={"ID":"6471b061-3dd3-46b3-ac86-9051b15355fa","Type":"ContainerDied","Data":"8c0acf8f9e452f2aff13aba0579aca8b5e810abce2f4d513c1d7f4760ddf5690"} Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.270069 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8fd57975-z5nmg" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.270918 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-t9jv2" event={"ID":"744dc616-99e5-4b30-bd16-476de000e0ae","Type":"ContainerStarted","Data":"203bff045b8dce96ed294acff9cb6426dc480b7f5dd5a0960c17cd5f5be38f7c"} Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.272478 4744 generic.go:334] "Generic (PLEG): container finished" podID="1d23fe19-d889-4aa8-9cc9-35eecb6449bc" containerID="e65f207bcf2f14ccb82b865966626dc284f7f15e57faebc3d79703a496df6b10" exitCode=0 Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.272507 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-rdzjs" event={"ID":"1d23fe19-d889-4aa8-9cc9-35eecb6449bc","Type":"ContainerDied","Data":"e65f207bcf2f14ccb82b865966626dc284f7f15e57faebc3d79703a496df6b10"} Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.272522 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-rdzjs" event={"ID":"1d23fe19-d889-4aa8-9cc9-35eecb6449bc","Type":"ContainerStarted","Data":"dad028de5dec19fc9b1ffd95f8036154afb6ddb2ba6e938b6464cf8b0e7764e4"} Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.309763 4744 scope.go:117] "RemoveContainer" containerID="2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.338097 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-z5nmg"] Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.349541 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d8fd57975-z5nmg"] Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.359885 4744 scope.go:117] "RemoveContainer" containerID="7afee42ccc67c77d229b6d74efe4c006d82b31a7f036f401fd258ff844240111" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.383211 4744 scope.go:117] "RemoveContainer" containerID="2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c" Oct 08 09:27:20 crc kubenswrapper[4744]: E1008 09:27:20.383891 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c\": container with ID starting with 2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c not found: ID does not exist" containerID="2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.383921 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c"} err="failed to get container status \"2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c\": rpc error: code = NotFound desc = could not find container \"2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c\": container with ID starting with 2aa86d0bcc1811f78958ccbce6dfe7a7ad170f57a46a9e3844703092c0ee7a4c not found: ID does not exist" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.383945 4744 scope.go:117] "RemoveContainer" containerID="7afee42ccc67c77d229b6d74efe4c006d82b31a7f036f401fd258ff844240111" Oct 08 09:27:20 crc kubenswrapper[4744]: E1008 09:27:20.384127 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7afee42ccc67c77d229b6d74efe4c006d82b31a7f036f401fd258ff844240111\": container with ID starting with 7afee42ccc67c77d229b6d74efe4c006d82b31a7f036f401fd258ff844240111 not found: ID does not exist" containerID="7afee42ccc67c77d229b6d74efe4c006d82b31a7f036f401fd258ff844240111" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.384156 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7afee42ccc67c77d229b6d74efe4c006d82b31a7f036f401fd258ff844240111"} err="failed to get container status \"7afee42ccc67c77d229b6d74efe4c006d82b31a7f036f401fd258ff844240111\": rpc error: code = NotFound desc = could not find container \"7afee42ccc67c77d229b6d74efe4c006d82b31a7f036f401fd258ff844240111\": container with ID starting with 7afee42ccc67c77d229b6d74efe4c006d82b31a7f036f401fd258ff844240111 not found: ID does not exist" Oct 08 09:27:20 crc kubenswrapper[4744]: I1008 09:27:20.858164 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 09:27:21 crc kubenswrapper[4744]: I1008 09:27:21.292805 4744 generic.go:334] "Generic (PLEG): container finished" podID="744dc616-99e5-4b30-bd16-476de000e0ae" containerID="bebda48be52e4d4862462d1a9b8d02e52be152f93dd5281c01b88bec75eec8c4" exitCode=0 Oct 08 09:27:21 crc kubenswrapper[4744]: I1008 09:27:21.293464 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-t9jv2" event={"ID":"744dc616-99e5-4b30-bd16-476de000e0ae","Type":"ContainerDied","Data":"bebda48be52e4d4862462d1a9b8d02e52be152f93dd5281c01b88bec75eec8c4"} Oct 08 09:27:21 crc kubenswrapper[4744]: I1008 09:27:21.466399 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6471b061-3dd3-46b3-ac86-9051b15355fa" path="/var/lib/kubelet/pods/6471b061-3dd3-46b3-ac86-9051b15355fa/volumes" Oct 08 09:27:21 crc kubenswrapper[4744]: I1008 09:27:21.679469 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-rdzjs" Oct 08 09:27:21 crc kubenswrapper[4744]: I1008 09:27:21.744865 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nklg6\" (UniqueName: \"kubernetes.io/projected/1d23fe19-d889-4aa8-9cc9-35eecb6449bc-kube-api-access-nklg6\") pod \"1d23fe19-d889-4aa8-9cc9-35eecb6449bc\" (UID: \"1d23fe19-d889-4aa8-9cc9-35eecb6449bc\") " Oct 08 09:27:21 crc kubenswrapper[4744]: I1008 09:27:21.762720 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d23fe19-d889-4aa8-9cc9-35eecb6449bc-kube-api-access-nklg6" (OuterVolumeSpecName: "kube-api-access-nklg6") pod "1d23fe19-d889-4aa8-9cc9-35eecb6449bc" (UID: "1d23fe19-d889-4aa8-9cc9-35eecb6449bc"). InnerVolumeSpecName "kube-api-access-nklg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:27:21 crc kubenswrapper[4744]: I1008 09:27:21.847312 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nklg6\" (UniqueName: \"kubernetes.io/projected/1d23fe19-d889-4aa8-9cc9-35eecb6449bc-kube-api-access-nklg6\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:22 crc kubenswrapper[4744]: I1008 09:27:22.311555 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-rdzjs" Oct 08 09:27:22 crc kubenswrapper[4744]: I1008 09:27:22.315384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-rdzjs" event={"ID":"1d23fe19-d889-4aa8-9cc9-35eecb6449bc","Type":"ContainerDied","Data":"dad028de5dec19fc9b1ffd95f8036154afb6ddb2ba6e938b6464cf8b0e7764e4"} Oct 08 09:27:22 crc kubenswrapper[4744]: I1008 09:27:22.315419 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dad028de5dec19fc9b1ffd95f8036154afb6ddb2ba6e938b6464cf8b0e7764e4" Oct 08 09:27:22 crc kubenswrapper[4744]: I1008 09:27:22.742646 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-t9jv2" Oct 08 09:27:22 crc kubenswrapper[4744]: I1008 09:27:22.864172 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6brcm\" (UniqueName: \"kubernetes.io/projected/744dc616-99e5-4b30-bd16-476de000e0ae-kube-api-access-6brcm\") pod \"744dc616-99e5-4b30-bd16-476de000e0ae\" (UID: \"744dc616-99e5-4b30-bd16-476de000e0ae\") " Oct 08 09:27:22 crc kubenswrapper[4744]: I1008 09:27:22.873623 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/744dc616-99e5-4b30-bd16-476de000e0ae-kube-api-access-6brcm" (OuterVolumeSpecName: "kube-api-access-6brcm") pod "744dc616-99e5-4b30-bd16-476de000e0ae" (UID: "744dc616-99e5-4b30-bd16-476de000e0ae"). InnerVolumeSpecName "kube-api-access-6brcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:27:22 crc kubenswrapper[4744]: I1008 09:27:22.966895 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6brcm\" (UniqueName: \"kubernetes.io/projected/744dc616-99e5-4b30-bd16-476de000e0ae-kube-api-access-6brcm\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:23 crc kubenswrapper[4744]: I1008 09:27:23.326396 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-t9jv2" event={"ID":"744dc616-99e5-4b30-bd16-476de000e0ae","Type":"ContainerDied","Data":"203bff045b8dce96ed294acff9cb6426dc480b7f5dd5a0960c17cd5f5be38f7c"} Oct 08 09:27:23 crc kubenswrapper[4744]: I1008 09:27:23.327472 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="203bff045b8dce96ed294acff9cb6426dc480b7f5dd5a0960c17cd5f5be38f7c" Oct 08 09:27:23 crc kubenswrapper[4744]: I1008 09:27:23.326522 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-t9jv2" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.445677 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b2c5-account-create-lhn7c"] Oct 08 09:27:24 crc kubenswrapper[4744]: E1008 09:27:24.446228 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d23fe19-d889-4aa8-9cc9-35eecb6449bc" containerName="mariadb-database-create" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.446247 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d23fe19-d889-4aa8-9cc9-35eecb6449bc" containerName="mariadb-database-create" Oct 08 09:27:24 crc kubenswrapper[4744]: E1008 09:27:24.446269 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="744dc616-99e5-4b30-bd16-476de000e0ae" containerName="mariadb-database-create" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.446278 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="744dc616-99e5-4b30-bd16-476de000e0ae" containerName="mariadb-database-create" Oct 08 09:27:24 crc kubenswrapper[4744]: E1008 09:27:24.446294 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6471b061-3dd3-46b3-ac86-9051b15355fa" containerName="init" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.446304 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6471b061-3dd3-46b3-ac86-9051b15355fa" containerName="init" Oct 08 09:27:24 crc kubenswrapper[4744]: E1008 09:27:24.446322 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6471b061-3dd3-46b3-ac86-9051b15355fa" containerName="dnsmasq-dns" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.446330 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6471b061-3dd3-46b3-ac86-9051b15355fa" containerName="dnsmasq-dns" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.446591 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d23fe19-d889-4aa8-9cc9-35eecb6449bc" containerName="mariadb-database-create" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.446610 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6471b061-3dd3-46b3-ac86-9051b15355fa" containerName="dnsmasq-dns" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.446622 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="744dc616-99e5-4b30-bd16-476de000e0ae" containerName="mariadb-database-create" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.447341 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b2c5-account-create-lhn7c" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.454026 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.459328 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b2c5-account-create-lhn7c"] Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.596010 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlkj8\" (UniqueName: \"kubernetes.io/projected/00f05c86-12ae-4a9d-bc7c-8c4619a68028-kube-api-access-zlkj8\") pod \"glance-b2c5-account-create-lhn7c\" (UID: \"00f05c86-12ae-4a9d-bc7c-8c4619a68028\") " pod="openstack/glance-b2c5-account-create-lhn7c" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.698184 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlkj8\" (UniqueName: \"kubernetes.io/projected/00f05c86-12ae-4a9d-bc7c-8c4619a68028-kube-api-access-zlkj8\") pod \"glance-b2c5-account-create-lhn7c\" (UID: \"00f05c86-12ae-4a9d-bc7c-8c4619a68028\") " pod="openstack/glance-b2c5-account-create-lhn7c" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.726804 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlkj8\" (UniqueName: \"kubernetes.io/projected/00f05c86-12ae-4a9d-bc7c-8c4619a68028-kube-api-access-zlkj8\") pod \"glance-b2c5-account-create-lhn7c\" (UID: \"00f05c86-12ae-4a9d-bc7c-8c4619a68028\") " pod="openstack/glance-b2c5-account-create-lhn7c" Oct 08 09:27:24 crc kubenswrapper[4744]: I1008 09:27:24.768585 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b2c5-account-create-lhn7c" Oct 08 09:27:25 crc kubenswrapper[4744]: I1008 09:27:25.300651 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b2c5-account-create-lhn7c"] Oct 08 09:27:25 crc kubenswrapper[4744]: I1008 09:27:25.348009 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b2c5-account-create-lhn7c" event={"ID":"00f05c86-12ae-4a9d-bc7c-8c4619a68028","Type":"ContainerStarted","Data":"2c266fc7965fa47057dbf907b643ca74aaab4b65d9a0482cd117de586124d11c"} Oct 08 09:27:26 crc kubenswrapper[4744]: I1008 09:27:26.377465 4744 generic.go:334] "Generic (PLEG): container finished" podID="00f05c86-12ae-4a9d-bc7c-8c4619a68028" containerID="8b9cf5919c6fc5f275d0718804ff1e4bc1028d06cbd99118545f053b45af183c" exitCode=0 Oct 08 09:27:26 crc kubenswrapper[4744]: I1008 09:27:26.377956 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b2c5-account-create-lhn7c" event={"ID":"00f05c86-12ae-4a9d-bc7c-8c4619a68028","Type":"ContainerDied","Data":"8b9cf5919c6fc5f275d0718804ff1e4bc1028d06cbd99118545f053b45af183c"} Oct 08 09:27:27 crc kubenswrapper[4744]: I1008 09:27:27.707668 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b2c5-account-create-lhn7c" Oct 08 09:27:27 crc kubenswrapper[4744]: I1008 09:27:27.859180 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlkj8\" (UniqueName: \"kubernetes.io/projected/00f05c86-12ae-4a9d-bc7c-8c4619a68028-kube-api-access-zlkj8\") pod \"00f05c86-12ae-4a9d-bc7c-8c4619a68028\" (UID: \"00f05c86-12ae-4a9d-bc7c-8c4619a68028\") " Oct 08 09:27:27 crc kubenswrapper[4744]: I1008 09:27:27.872663 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00f05c86-12ae-4a9d-bc7c-8c4619a68028-kube-api-access-zlkj8" (OuterVolumeSpecName: "kube-api-access-zlkj8") pod "00f05c86-12ae-4a9d-bc7c-8c4619a68028" (UID: "00f05c86-12ae-4a9d-bc7c-8c4619a68028"). InnerVolumeSpecName "kube-api-access-zlkj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:27:27 crc kubenswrapper[4744]: I1008 09:27:27.961394 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlkj8\" (UniqueName: \"kubernetes.io/projected/00f05c86-12ae-4a9d-bc7c-8c4619a68028-kube-api-access-zlkj8\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:28 crc kubenswrapper[4744]: I1008 09:27:28.398008 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b2c5-account-create-lhn7c" event={"ID":"00f05c86-12ae-4a9d-bc7c-8c4619a68028","Type":"ContainerDied","Data":"2c266fc7965fa47057dbf907b643ca74aaab4b65d9a0482cd117de586124d11c"} Oct 08 09:27:28 crc kubenswrapper[4744]: I1008 09:27:28.398053 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c266fc7965fa47057dbf907b643ca74aaab4b65d9a0482cd117de586124d11c" Oct 08 09:27:28 crc kubenswrapper[4744]: I1008 09:27:28.398188 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b2c5-account-create-lhn7c" Oct 08 09:27:28 crc kubenswrapper[4744]: I1008 09:27:28.881741 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-a44f-account-create-qsjwh"] Oct 08 09:27:28 crc kubenswrapper[4744]: E1008 09:27:28.882132 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f05c86-12ae-4a9d-bc7c-8c4619a68028" containerName="mariadb-account-create" Oct 08 09:27:28 crc kubenswrapper[4744]: I1008 09:27:28.882147 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f05c86-12ae-4a9d-bc7c-8c4619a68028" containerName="mariadb-account-create" Oct 08 09:27:28 crc kubenswrapper[4744]: I1008 09:27:28.882317 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="00f05c86-12ae-4a9d-bc7c-8c4619a68028" containerName="mariadb-account-create" Oct 08 09:27:28 crc kubenswrapper[4744]: I1008 09:27:28.882924 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-a44f-account-create-qsjwh" Oct 08 09:27:28 crc kubenswrapper[4744]: I1008 09:27:28.885061 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 08 09:27:28 crc kubenswrapper[4744]: I1008 09:27:28.894361 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-a44f-account-create-qsjwh"] Oct 08 09:27:28 crc kubenswrapper[4744]: I1008 09:27:28.978644 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf5x5\" (UniqueName: \"kubernetes.io/projected/b5b99de0-ebdd-460b-a6e9-011911768480-kube-api-access-cf5x5\") pod \"keystone-a44f-account-create-qsjwh\" (UID: \"b5b99de0-ebdd-460b-a6e9-011911768480\") " pod="openstack/keystone-a44f-account-create-qsjwh" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.080617 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf5x5\" (UniqueName: \"kubernetes.io/projected/b5b99de0-ebdd-460b-a6e9-011911768480-kube-api-access-cf5x5\") pod \"keystone-a44f-account-create-qsjwh\" (UID: \"b5b99de0-ebdd-460b-a6e9-011911768480\") " pod="openstack/keystone-a44f-account-create-qsjwh" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.097313 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf5x5\" (UniqueName: \"kubernetes.io/projected/b5b99de0-ebdd-460b-a6e9-011911768480-kube-api-access-cf5x5\") pod \"keystone-a44f-account-create-qsjwh\" (UID: \"b5b99de0-ebdd-460b-a6e9-011911768480\") " pod="openstack/keystone-a44f-account-create-qsjwh" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.212200 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-a44f-account-create-qsjwh" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.345552 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f3a6-account-create-2cnwz"] Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.346700 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f3a6-account-create-2cnwz" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.350034 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.363602 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f3a6-account-create-2cnwz"] Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.489067 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rpts\" (UniqueName: \"kubernetes.io/projected/39dc9dd6-c3f6-4f28-b66a-3908b2585c5d-kube-api-access-8rpts\") pod \"placement-f3a6-account-create-2cnwz\" (UID: \"39dc9dd6-c3f6-4f28-b66a-3908b2585c5d\") " pod="openstack/placement-f3a6-account-create-2cnwz" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.572157 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-fdrfp"] Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.573152 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.575420 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lwftp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.575616 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.592598 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rpts\" (UniqueName: \"kubernetes.io/projected/39dc9dd6-c3f6-4f28-b66a-3908b2585c5d-kube-api-access-8rpts\") pod \"placement-f3a6-account-create-2cnwz\" (UID: \"39dc9dd6-c3f6-4f28-b66a-3908b2585c5d\") " pod="openstack/placement-f3a6-account-create-2cnwz" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.594152 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-fdrfp"] Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.623644 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rpts\" (UniqueName: \"kubernetes.io/projected/39dc9dd6-c3f6-4f28-b66a-3908b2585c5d-kube-api-access-8rpts\") pod \"placement-f3a6-account-create-2cnwz\" (UID: \"39dc9dd6-c3f6-4f28-b66a-3908b2585c5d\") " pod="openstack/placement-f3a6-account-create-2cnwz" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.674062 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f3a6-account-create-2cnwz" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.693947 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-combined-ca-bundle\") pod \"glance-db-sync-fdrfp\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.693998 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-config-data\") pod \"glance-db-sync-fdrfp\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.694024 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68cp5\" (UniqueName: \"kubernetes.io/projected/f33787b3-d366-41b9-80f7-955dc458cb93-kube-api-access-68cp5\") pod \"glance-db-sync-fdrfp\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.694056 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-db-sync-config-data\") pod \"glance-db-sync-fdrfp\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.728303 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-a44f-account-create-qsjwh"] Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.795442 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-combined-ca-bundle\") pod \"glance-db-sync-fdrfp\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.795713 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-config-data\") pod \"glance-db-sync-fdrfp\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.795740 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68cp5\" (UniqueName: \"kubernetes.io/projected/f33787b3-d366-41b9-80f7-955dc458cb93-kube-api-access-68cp5\") pod \"glance-db-sync-fdrfp\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.795773 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-db-sync-config-data\") pod \"glance-db-sync-fdrfp\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.801094 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-combined-ca-bundle\") pod \"glance-db-sync-fdrfp\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.802302 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-db-sync-config-data\") pod \"glance-db-sync-fdrfp\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.802698 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-config-data\") pod \"glance-db-sync-fdrfp\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.816435 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68cp5\" (UniqueName: \"kubernetes.io/projected/f33787b3-d366-41b9-80f7-955dc458cb93-kube-api-access-68cp5\") pod \"glance-db-sync-fdrfp\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:29 crc kubenswrapper[4744]: I1008 09:27:29.897541 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:30 crc kubenswrapper[4744]: I1008 09:27:30.099507 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 08 09:27:30 crc kubenswrapper[4744]: I1008 09:27:30.184037 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f3a6-account-create-2cnwz"] Oct 08 09:27:30 crc kubenswrapper[4744]: I1008 09:27:30.347648 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-fdrfp"] Oct 08 09:27:30 crc kubenswrapper[4744]: W1008 09:27:30.350843 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf33787b3_d366_41b9_80f7_955dc458cb93.slice/crio-a867a3066d9aad8fd8e6cfd163653f1078b13f2b22e0d0843b8ef8261ca956a6 WatchSource:0}: Error finding container a867a3066d9aad8fd8e6cfd163653f1078b13f2b22e0d0843b8ef8261ca956a6: Status 404 returned error can't find the container with id a867a3066d9aad8fd8e6cfd163653f1078b13f2b22e0d0843b8ef8261ca956a6 Oct 08 09:27:30 crc kubenswrapper[4744]: I1008 09:27:30.423004 4744 generic.go:334] "Generic (PLEG): container finished" podID="b5b99de0-ebdd-460b-a6e9-011911768480" containerID="3437fcc9c340f5c61a710ed66d4b69c51113ba5cd0bcd4197aa2b7ed20ba27e7" exitCode=0 Oct 08 09:27:30 crc kubenswrapper[4744]: I1008 09:27:30.423432 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-a44f-account-create-qsjwh" event={"ID":"b5b99de0-ebdd-460b-a6e9-011911768480","Type":"ContainerDied","Data":"3437fcc9c340f5c61a710ed66d4b69c51113ba5cd0bcd4197aa2b7ed20ba27e7"} Oct 08 09:27:30 crc kubenswrapper[4744]: I1008 09:27:30.423546 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-a44f-account-create-qsjwh" event={"ID":"b5b99de0-ebdd-460b-a6e9-011911768480","Type":"ContainerStarted","Data":"d78cdd6502af0a847ede80cbbb758438492aaa17714f9de47a869859a917cd78"} Oct 08 09:27:30 crc kubenswrapper[4744]: I1008 09:27:30.429599 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f3a6-account-create-2cnwz" event={"ID":"39dc9dd6-c3f6-4f28-b66a-3908b2585c5d","Type":"ContainerStarted","Data":"d41c9bde20394444eb92687c78b780430aee1a63b3b47f331802df7fc1d0421b"} Oct 08 09:27:30 crc kubenswrapper[4744]: I1008 09:27:30.432002 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fdrfp" event={"ID":"f33787b3-d366-41b9-80f7-955dc458cb93","Type":"ContainerStarted","Data":"a867a3066d9aad8fd8e6cfd163653f1078b13f2b22e0d0843b8ef8261ca956a6"} Oct 08 09:27:31 crc kubenswrapper[4744]: I1008 09:27:31.444234 4744 generic.go:334] "Generic (PLEG): container finished" podID="39dc9dd6-c3f6-4f28-b66a-3908b2585c5d" containerID="c76c311a1bafdd7b9488148f76accf280f866bf3c9bc6f06367c38f3dbf50b6b" exitCode=0 Oct 08 09:27:31 crc kubenswrapper[4744]: I1008 09:27:31.444334 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f3a6-account-create-2cnwz" event={"ID":"39dc9dd6-c3f6-4f28-b66a-3908b2585c5d","Type":"ContainerDied","Data":"c76c311a1bafdd7b9488148f76accf280f866bf3c9bc6f06367c38f3dbf50b6b"} Oct 08 09:27:31 crc kubenswrapper[4744]: I1008 09:27:31.766627 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-a44f-account-create-qsjwh" Oct 08 09:27:31 crc kubenswrapper[4744]: I1008 09:27:31.838897 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf5x5\" (UniqueName: \"kubernetes.io/projected/b5b99de0-ebdd-460b-a6e9-011911768480-kube-api-access-cf5x5\") pod \"b5b99de0-ebdd-460b-a6e9-011911768480\" (UID: \"b5b99de0-ebdd-460b-a6e9-011911768480\") " Oct 08 09:27:31 crc kubenswrapper[4744]: I1008 09:27:31.874344 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b99de0-ebdd-460b-a6e9-011911768480-kube-api-access-cf5x5" (OuterVolumeSpecName: "kube-api-access-cf5x5") pod "b5b99de0-ebdd-460b-a6e9-011911768480" (UID: "b5b99de0-ebdd-460b-a6e9-011911768480"). InnerVolumeSpecName "kube-api-access-cf5x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:27:31 crc kubenswrapper[4744]: I1008 09:27:31.941019 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf5x5\" (UniqueName: \"kubernetes.io/projected/b5b99de0-ebdd-460b-a6e9-011911768480-kube-api-access-cf5x5\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:32 crc kubenswrapper[4744]: I1008 09:27:32.456627 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-a44f-account-create-qsjwh" event={"ID":"b5b99de0-ebdd-460b-a6e9-011911768480","Type":"ContainerDied","Data":"d78cdd6502af0a847ede80cbbb758438492aaa17714f9de47a869859a917cd78"} Oct 08 09:27:32 crc kubenswrapper[4744]: I1008 09:27:32.457124 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d78cdd6502af0a847ede80cbbb758438492aaa17714f9de47a869859a917cd78" Oct 08 09:27:32 crc kubenswrapper[4744]: I1008 09:27:32.456678 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-a44f-account-create-qsjwh" Oct 08 09:27:32 crc kubenswrapper[4744]: I1008 09:27:32.785116 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f3a6-account-create-2cnwz" Oct 08 09:27:32 crc kubenswrapper[4744]: I1008 09:27:32.855836 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rpts\" (UniqueName: \"kubernetes.io/projected/39dc9dd6-c3f6-4f28-b66a-3908b2585c5d-kube-api-access-8rpts\") pod \"39dc9dd6-c3f6-4f28-b66a-3908b2585c5d\" (UID: \"39dc9dd6-c3f6-4f28-b66a-3908b2585c5d\") " Oct 08 09:27:32 crc kubenswrapper[4744]: I1008 09:27:32.864896 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39dc9dd6-c3f6-4f28-b66a-3908b2585c5d-kube-api-access-8rpts" (OuterVolumeSpecName: "kube-api-access-8rpts") pod "39dc9dd6-c3f6-4f28-b66a-3908b2585c5d" (UID: "39dc9dd6-c3f6-4f28-b66a-3908b2585c5d"). InnerVolumeSpecName "kube-api-access-8rpts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:27:32 crc kubenswrapper[4744]: I1008 09:27:32.958520 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rpts\" (UniqueName: \"kubernetes.io/projected/39dc9dd6-c3f6-4f28-b66a-3908b2585c5d-kube-api-access-8rpts\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:33 crc kubenswrapper[4744]: I1008 09:27:33.476876 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f3a6-account-create-2cnwz" Oct 08 09:27:33 crc kubenswrapper[4744]: I1008 09:27:33.498732 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f3a6-account-create-2cnwz" event={"ID":"39dc9dd6-c3f6-4f28-b66a-3908b2585c5d","Type":"ContainerDied","Data":"d41c9bde20394444eb92687c78b780430aee1a63b3b47f331802df7fc1d0421b"} Oct 08 09:27:33 crc kubenswrapper[4744]: I1008 09:27:33.498796 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d41c9bde20394444eb92687c78b780430aee1a63b3b47f331802df7fc1d0421b" Oct 08 09:27:39 crc kubenswrapper[4744]: I1008 09:27:39.794409 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-mhqr7" podUID="ad8606dc-c6d0-4ae7-905e-929d0a82ae8d" containerName="ovn-controller" probeResult="failure" output=< Oct 08 09:27:39 crc kubenswrapper[4744]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 08 09:27:39 crc kubenswrapper[4744]: > Oct 08 09:27:41 crc kubenswrapper[4744]: I1008 09:27:41.558572 4744 generic.go:334] "Generic (PLEG): container finished" podID="18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" containerID="a7ae4524f517636f791ec2332601150d328e8d3eb0890dfd3c61265b8ed54e20" exitCode=0 Oct 08 09:27:41 crc kubenswrapper[4744]: I1008 09:27:41.558661 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6","Type":"ContainerDied","Data":"a7ae4524f517636f791ec2332601150d328e8d3eb0890dfd3c61265b8ed54e20"} Oct 08 09:27:41 crc kubenswrapper[4744]: I1008 09:27:41.562729 4744 generic.go:334] "Generic (PLEG): container finished" podID="6a6f5a1b-6abb-4568-89b5-a255f919cebb" containerID="816a11709d28ae06d2f534850a7bff8d19e598d0723d36416765dd4b97d1863e" exitCode=0 Oct 08 09:27:41 crc kubenswrapper[4744]: I1008 09:27:41.562801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6a6f5a1b-6abb-4568-89b5-a255f919cebb","Type":"ContainerDied","Data":"816a11709d28ae06d2f534850a7bff8d19e598d0723d36416765dd4b97d1863e"} Oct 08 09:27:44 crc kubenswrapper[4744]: I1008 09:27:44.597786 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6a6f5a1b-6abb-4568-89b5-a255f919cebb","Type":"ContainerStarted","Data":"ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72"} Oct 08 09:27:44 crc kubenswrapper[4744]: I1008 09:27:44.598818 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 08 09:27:44 crc kubenswrapper[4744]: I1008 09:27:44.600558 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fdrfp" event={"ID":"f33787b3-d366-41b9-80f7-955dc458cb93","Type":"ContainerStarted","Data":"6ef8d9d17f46150fccc932f1a7f082a31010389556aa962745653259a5f12e5d"} Oct 08 09:27:44 crc kubenswrapper[4744]: I1008 09:27:44.604688 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6","Type":"ContainerStarted","Data":"f3cb2f38c332b38d6da190637c53552d6176707980c4a0f214e672ede5b5273e"} Oct 08 09:27:44 crc kubenswrapper[4744]: I1008 09:27:44.605036 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:27:44 crc kubenswrapper[4744]: I1008 09:27:44.632916 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.878767971 podStartE2EDuration="1m20.632889737s" podCreationTimestamp="2025-10-08 09:26:24 +0000 UTC" firstStartedPulling="2025-10-08 09:26:26.813089069 +0000 UTC m=+882.060734308" lastFinishedPulling="2025-10-08 09:27:07.567210835 +0000 UTC m=+922.814856074" observedRunningTime="2025-10-08 09:27:44.625085844 +0000 UTC m=+959.872731123" watchObservedRunningTime="2025-10-08 09:27:44.632889737 +0000 UTC m=+959.880534996" Oct 08 09:27:44 crc kubenswrapper[4744]: I1008 09:27:44.654290 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-fdrfp" podStartSLOduration=2.3458485700000002 podStartE2EDuration="15.654271876s" podCreationTimestamp="2025-10-08 09:27:29 +0000 UTC" firstStartedPulling="2025-10-08 09:27:30.353666143 +0000 UTC m=+945.601311382" lastFinishedPulling="2025-10-08 09:27:43.662089449 +0000 UTC m=+958.909734688" observedRunningTime="2025-10-08 09:27:44.646890486 +0000 UTC m=+959.894535735" watchObservedRunningTime="2025-10-08 09:27:44.654271876 +0000 UTC m=+959.901917115" Oct 08 09:27:44 crc kubenswrapper[4744]: I1008 09:27:44.681967 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.213798379 podStartE2EDuration="1m20.681917214s" podCreationTimestamp="2025-10-08 09:26:24 +0000 UTC" firstStartedPulling="2025-10-08 09:26:27.037666569 +0000 UTC m=+882.285311808" lastFinishedPulling="2025-10-08 09:27:07.505785404 +0000 UTC m=+922.753430643" observedRunningTime="2025-10-08 09:27:44.674672198 +0000 UTC m=+959.922317457" watchObservedRunningTime="2025-10-08 09:27:44.681917214 +0000 UTC m=+959.929562453" Oct 08 09:27:44 crc kubenswrapper[4744]: I1008 09:27:44.813810 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:27:44 crc kubenswrapper[4744]: I1008 09:27:44.828627 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-ff8dm" Oct 08 09:27:44 crc kubenswrapper[4744]: I1008 09:27:44.845286 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-mhqr7" podUID="ad8606dc-c6d0-4ae7-905e-929d0a82ae8d" containerName="ovn-controller" probeResult="failure" output=< Oct 08 09:27:44 crc kubenswrapper[4744]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 08 09:27:44 crc kubenswrapper[4744]: > Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.071789 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-mhqr7-config-4g2zc"] Oct 08 09:27:45 crc kubenswrapper[4744]: E1008 09:27:45.072102 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39dc9dd6-c3f6-4f28-b66a-3908b2585c5d" containerName="mariadb-account-create" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.072120 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="39dc9dd6-c3f6-4f28-b66a-3908b2585c5d" containerName="mariadb-account-create" Oct 08 09:27:45 crc kubenswrapper[4744]: E1008 09:27:45.072152 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b99de0-ebdd-460b-a6e9-011911768480" containerName="mariadb-account-create" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.072159 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b99de0-ebdd-460b-a6e9-011911768480" containerName="mariadb-account-create" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.072327 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="39dc9dd6-c3f6-4f28-b66a-3908b2585c5d" containerName="mariadb-account-create" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.072359 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b99de0-ebdd-460b-a6e9-011911768480" containerName="mariadb-account-create" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.072850 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.078390 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.085390 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mhqr7-config-4g2zc"] Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.216599 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-run\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.216663 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-log-ovn\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.216682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c43f9c29-354c-405c-bada-395172934d9d-scripts\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.216706 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c43f9c29-354c-405c-bada-395172934d9d-additional-scripts\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.216923 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-run-ovn\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.217005 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8bhv\" (UniqueName: \"kubernetes.io/projected/c43f9c29-354c-405c-bada-395172934d9d-kube-api-access-m8bhv\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.318255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c43f9c29-354c-405c-bada-395172934d9d-additional-scripts\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.318444 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-run-ovn\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.318487 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8bhv\" (UniqueName: \"kubernetes.io/projected/c43f9c29-354c-405c-bada-395172934d9d-kube-api-access-m8bhv\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.318549 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-run\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.318600 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-log-ovn\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.318631 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c43f9c29-354c-405c-bada-395172934d9d-scripts\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.318766 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-run-ovn\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.318839 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-run\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.318972 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c43f9c29-354c-405c-bada-395172934d9d-additional-scripts\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.319034 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-log-ovn\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.322343 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c43f9c29-354c-405c-bada-395172934d9d-scripts\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.345004 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8bhv\" (UniqueName: \"kubernetes.io/projected/c43f9c29-354c-405c-bada-395172934d9d-kube-api-access-m8bhv\") pod \"ovn-controller-mhqr7-config-4g2zc\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.392543 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:45 crc kubenswrapper[4744]: I1008 09:27:45.843905 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mhqr7-config-4g2zc"] Oct 08 09:27:46 crc kubenswrapper[4744]: I1008 09:27:46.629830 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mhqr7-config-4g2zc" event={"ID":"c43f9c29-354c-405c-bada-395172934d9d","Type":"ContainerStarted","Data":"495bda41d442a59fe8cd6589342d3703bda62335e287e8c20795bdcb2771766d"} Oct 08 09:27:48 crc kubenswrapper[4744]: I1008 09:27:48.649073 4744 generic.go:334] "Generic (PLEG): container finished" podID="c43f9c29-354c-405c-bada-395172934d9d" containerID="02160fdf25f738b58ca129e0f0a8927722ca3ba72b1ea0fc0371d083ba9304d5" exitCode=0 Oct 08 09:27:48 crc kubenswrapper[4744]: I1008 09:27:48.649113 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mhqr7-config-4g2zc" event={"ID":"c43f9c29-354c-405c-bada-395172934d9d","Type":"ContainerDied","Data":"02160fdf25f738b58ca129e0f0a8927722ca3ba72b1ea0fc0371d083ba9304d5"} Oct 08 09:27:49 crc kubenswrapper[4744]: I1008 09:27:49.825078 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-mhqr7" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:49.999734 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.117984 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-run-ovn\") pod \"c43f9c29-354c-405c-bada-395172934d9d\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.118091 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-log-ovn\") pod \"c43f9c29-354c-405c-bada-395172934d9d\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.118138 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8bhv\" (UniqueName: \"kubernetes.io/projected/c43f9c29-354c-405c-bada-395172934d9d-kube-api-access-m8bhv\") pod \"c43f9c29-354c-405c-bada-395172934d9d\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.118128 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c43f9c29-354c-405c-bada-395172934d9d" (UID: "c43f9c29-354c-405c-bada-395172934d9d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.118223 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c43f9c29-354c-405c-bada-395172934d9d-additional-scripts\") pod \"c43f9c29-354c-405c-bada-395172934d9d\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.118303 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c43f9c29-354c-405c-bada-395172934d9d-scripts\") pod \"c43f9c29-354c-405c-bada-395172934d9d\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.118160 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c43f9c29-354c-405c-bada-395172934d9d" (UID: "c43f9c29-354c-405c-bada-395172934d9d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.118390 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-run\") pod \"c43f9c29-354c-405c-bada-395172934d9d\" (UID: \"c43f9c29-354c-405c-bada-395172934d9d\") " Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.118530 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-run" (OuterVolumeSpecName: "var-run") pod "c43f9c29-354c-405c-bada-395172934d9d" (UID: "c43f9c29-354c-405c-bada-395172934d9d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.118772 4744 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.118789 4744 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.118802 4744 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c43f9c29-354c-405c-bada-395172934d9d-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.118830 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c43f9c29-354c-405c-bada-395172934d9d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "c43f9c29-354c-405c-bada-395172934d9d" (UID: "c43f9c29-354c-405c-bada-395172934d9d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.119050 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c43f9c29-354c-405c-bada-395172934d9d-scripts" (OuterVolumeSpecName: "scripts") pod "c43f9c29-354c-405c-bada-395172934d9d" (UID: "c43f9c29-354c-405c-bada-395172934d9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.136823 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c43f9c29-354c-405c-bada-395172934d9d-kube-api-access-m8bhv" (OuterVolumeSpecName: "kube-api-access-m8bhv") pod "c43f9c29-354c-405c-bada-395172934d9d" (UID: "c43f9c29-354c-405c-bada-395172934d9d"). InnerVolumeSpecName "kube-api-access-m8bhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.220270 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8bhv\" (UniqueName: \"kubernetes.io/projected/c43f9c29-354c-405c-bada-395172934d9d-kube-api-access-m8bhv\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.220306 4744 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c43f9c29-354c-405c-bada-395172934d9d-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.220317 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c43f9c29-354c-405c-bada-395172934d9d-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.667927 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mhqr7-config-4g2zc" event={"ID":"c43f9c29-354c-405c-bada-395172934d9d","Type":"ContainerDied","Data":"495bda41d442a59fe8cd6589342d3703bda62335e287e8c20795bdcb2771766d"} Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.668017 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="495bda41d442a59fe8cd6589342d3703bda62335e287e8c20795bdcb2771766d" Oct 08 09:27:50 crc kubenswrapper[4744]: I1008 09:27:50.668126 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mhqr7-config-4g2zc" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.121701 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-mhqr7-config-4g2zc"] Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.130011 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-mhqr7-config-4g2zc"] Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.222951 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-mhqr7-config-w4wzk"] Oct 08 09:27:51 crc kubenswrapper[4744]: E1008 09:27:51.223255 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c43f9c29-354c-405c-bada-395172934d9d" containerName="ovn-config" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.223266 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c43f9c29-354c-405c-bada-395172934d9d" containerName="ovn-config" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.223456 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c43f9c29-354c-405c-bada-395172934d9d" containerName="ovn-config" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.223952 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.230040 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.246659 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mhqr7-config-w4wzk"] Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.341155 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21399e84-99fe-41b2-be88-fa30eb0f4246-scripts\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.341220 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-log-ovn\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.341292 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9wdr\" (UniqueName: \"kubernetes.io/projected/21399e84-99fe-41b2-be88-fa30eb0f4246-kube-api-access-j9wdr\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.341316 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-run\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.341348 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/21399e84-99fe-41b2-be88-fa30eb0f4246-additional-scripts\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.341426 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-run-ovn\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.443261 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-run-ovn\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.443331 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21399e84-99fe-41b2-be88-fa30eb0f4246-scripts\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.443385 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-log-ovn\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.443433 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9wdr\" (UniqueName: \"kubernetes.io/projected/21399e84-99fe-41b2-be88-fa30eb0f4246-kube-api-access-j9wdr\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.443451 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-run\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.443475 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/21399e84-99fe-41b2-be88-fa30eb0f4246-additional-scripts\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.443811 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-log-ovn\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.443885 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-run-ovn\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.444185 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/21399e84-99fe-41b2-be88-fa30eb0f4246-additional-scripts\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.444818 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-run\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.445739 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21399e84-99fe-41b2-be88-fa30eb0f4246-scripts\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.461972 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c43f9c29-354c-405c-bada-395172934d9d" path="/var/lib/kubelet/pods/c43f9c29-354c-405c-bada-395172934d9d/volumes" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.479416 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9wdr\" (UniqueName: \"kubernetes.io/projected/21399e84-99fe-41b2-be88-fa30eb0f4246-kube-api-access-j9wdr\") pod \"ovn-controller-mhqr7-config-w4wzk\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.541200 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:51 crc kubenswrapper[4744]: I1008 09:27:51.901643 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mhqr7-config-w4wzk"] Oct 08 09:27:52 crc kubenswrapper[4744]: I1008 09:27:52.696073 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mhqr7-config-w4wzk" event={"ID":"21399e84-99fe-41b2-be88-fa30eb0f4246","Type":"ContainerStarted","Data":"2fd80a6c6a013ebf4855e67c1c9e44f84a6e75f18ffcae45276348043008e93c"} Oct 08 09:27:52 crc kubenswrapper[4744]: I1008 09:27:52.696579 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mhqr7-config-w4wzk" event={"ID":"21399e84-99fe-41b2-be88-fa30eb0f4246","Type":"ContainerStarted","Data":"0bc2418f46657e0c4defc45a86c807b1d61c8fd3d9708bc77564905fa16c3023"} Oct 08 09:27:52 crc kubenswrapper[4744]: I1008 09:27:52.714731 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-mhqr7-config-w4wzk" podStartSLOduration=1.7146994850000001 podStartE2EDuration="1.714699485s" podCreationTimestamp="2025-10-08 09:27:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:27:52.713085638 +0000 UTC m=+967.960730877" watchObservedRunningTime="2025-10-08 09:27:52.714699485 +0000 UTC m=+967.962344724" Oct 08 09:27:53 crc kubenswrapper[4744]: I1008 09:27:53.707830 4744 generic.go:334] "Generic (PLEG): container finished" podID="21399e84-99fe-41b2-be88-fa30eb0f4246" containerID="2fd80a6c6a013ebf4855e67c1c9e44f84a6e75f18ffcae45276348043008e93c" exitCode=0 Oct 08 09:27:53 crc kubenswrapper[4744]: I1008 09:27:53.707974 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mhqr7-config-w4wzk" event={"ID":"21399e84-99fe-41b2-be88-fa30eb0f4246","Type":"ContainerDied","Data":"2fd80a6c6a013ebf4855e67c1c9e44f84a6e75f18ffcae45276348043008e93c"} Oct 08 09:27:54 crc kubenswrapper[4744]: I1008 09:27:54.717633 4744 generic.go:334] "Generic (PLEG): container finished" podID="f33787b3-d366-41b9-80f7-955dc458cb93" containerID="6ef8d9d17f46150fccc932f1a7f082a31010389556aa962745653259a5f12e5d" exitCode=0 Oct 08 09:27:54 crc kubenswrapper[4744]: I1008 09:27:54.717871 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fdrfp" event={"ID":"f33787b3-d366-41b9-80f7-955dc458cb93","Type":"ContainerDied","Data":"6ef8d9d17f46150fccc932f1a7f082a31010389556aa962745653259a5f12e5d"} Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.095949 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.208120 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-run-ovn\") pod \"21399e84-99fe-41b2-be88-fa30eb0f4246\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.208199 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-log-ovn\") pod \"21399e84-99fe-41b2-be88-fa30eb0f4246\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.208233 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "21399e84-99fe-41b2-be88-fa30eb0f4246" (UID: "21399e84-99fe-41b2-be88-fa30eb0f4246"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.208276 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9wdr\" (UniqueName: \"kubernetes.io/projected/21399e84-99fe-41b2-be88-fa30eb0f4246-kube-api-access-j9wdr\") pod \"21399e84-99fe-41b2-be88-fa30eb0f4246\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.208271 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "21399e84-99fe-41b2-be88-fa30eb0f4246" (UID: "21399e84-99fe-41b2-be88-fa30eb0f4246"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.208329 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-run\") pod \"21399e84-99fe-41b2-be88-fa30eb0f4246\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.208407 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-run" (OuterVolumeSpecName: "var-run") pod "21399e84-99fe-41b2-be88-fa30eb0f4246" (UID: "21399e84-99fe-41b2-be88-fa30eb0f4246"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.208483 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/21399e84-99fe-41b2-be88-fa30eb0f4246-additional-scripts\") pod \"21399e84-99fe-41b2-be88-fa30eb0f4246\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.208518 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21399e84-99fe-41b2-be88-fa30eb0f4246-scripts\") pod \"21399e84-99fe-41b2-be88-fa30eb0f4246\" (UID: \"21399e84-99fe-41b2-be88-fa30eb0f4246\") " Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.208868 4744 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.208888 4744 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.208899 4744 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/21399e84-99fe-41b2-be88-fa30eb0f4246-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.209233 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21399e84-99fe-41b2-be88-fa30eb0f4246-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "21399e84-99fe-41b2-be88-fa30eb0f4246" (UID: "21399e84-99fe-41b2-be88-fa30eb0f4246"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.209436 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21399e84-99fe-41b2-be88-fa30eb0f4246-scripts" (OuterVolumeSpecName: "scripts") pod "21399e84-99fe-41b2-be88-fa30eb0f4246" (UID: "21399e84-99fe-41b2-be88-fa30eb0f4246"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.215120 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21399e84-99fe-41b2-be88-fa30eb0f4246-kube-api-access-j9wdr" (OuterVolumeSpecName: "kube-api-access-j9wdr") pod "21399e84-99fe-41b2-be88-fa30eb0f4246" (UID: "21399e84-99fe-41b2-be88-fa30eb0f4246"). InnerVolumeSpecName "kube-api-access-j9wdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.309892 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9wdr\" (UniqueName: \"kubernetes.io/projected/21399e84-99fe-41b2-be88-fa30eb0f4246-kube-api-access-j9wdr\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.310167 4744 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/21399e84-99fe-41b2-be88-fa30eb0f4246-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.310237 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/21399e84-99fe-41b2-be88-fa30eb0f4246-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.659211 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="6a6f5a1b-6abb-4568-89b5-a255f919cebb" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.727811 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mhqr7-config-w4wzk" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.727799 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mhqr7-config-w4wzk" event={"ID":"21399e84-99fe-41b2-be88-fa30eb0f4246","Type":"ContainerDied","Data":"0bc2418f46657e0c4defc45a86c807b1d61c8fd3d9708bc77564905fa16c3023"} Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.730309 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bc2418f46657e0c4defc45a86c807b1d61c8fd3d9708bc77564905fa16c3023" Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.855193 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-mhqr7-config-w4wzk"] Oct 08 09:27:55 crc kubenswrapper[4744]: I1008 09:27:55.866746 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-mhqr7-config-w4wzk"] Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.087498 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.479944 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.535741 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-combined-ca-bundle\") pod \"f33787b3-d366-41b9-80f7-955dc458cb93\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.535826 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68cp5\" (UniqueName: \"kubernetes.io/projected/f33787b3-d366-41b9-80f7-955dc458cb93-kube-api-access-68cp5\") pod \"f33787b3-d366-41b9-80f7-955dc458cb93\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.535907 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-db-sync-config-data\") pod \"f33787b3-d366-41b9-80f7-955dc458cb93\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.535933 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-config-data\") pod \"f33787b3-d366-41b9-80f7-955dc458cb93\" (UID: \"f33787b3-d366-41b9-80f7-955dc458cb93\") " Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.545994 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f33787b3-d366-41b9-80f7-955dc458cb93-kube-api-access-68cp5" (OuterVolumeSpecName: "kube-api-access-68cp5") pod "f33787b3-d366-41b9-80f7-955dc458cb93" (UID: "f33787b3-d366-41b9-80f7-955dc458cb93"). InnerVolumeSpecName "kube-api-access-68cp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.558166 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f33787b3-d366-41b9-80f7-955dc458cb93" (UID: "f33787b3-d366-41b9-80f7-955dc458cb93"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.566097 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f33787b3-d366-41b9-80f7-955dc458cb93" (UID: "f33787b3-d366-41b9-80f7-955dc458cb93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.638848 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.638878 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68cp5\" (UniqueName: \"kubernetes.io/projected/f33787b3-d366-41b9-80f7-955dc458cb93-kube-api-access-68cp5\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.638891 4744 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.644389 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-config-data" (OuterVolumeSpecName: "config-data") pod "f33787b3-d366-41b9-80f7-955dc458cb93" (UID: "f33787b3-d366-41b9-80f7-955dc458cb93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.736636 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fdrfp" event={"ID":"f33787b3-d366-41b9-80f7-955dc458cb93","Type":"ContainerDied","Data":"a867a3066d9aad8fd8e6cfd163653f1078b13f2b22e0d0843b8ef8261ca956a6"} Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.736683 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a867a3066d9aad8fd8e6cfd163653f1078b13f2b22e0d0843b8ef8261ca956a6" Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.736746 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fdrfp" Oct 08 09:27:56 crc kubenswrapper[4744]: I1008 09:27:56.740593 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f33787b3-d366-41b9-80f7-955dc458cb93-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.305985 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq"] Oct 08 09:27:57 crc kubenswrapper[4744]: E1008 09:27:57.306574 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21399e84-99fe-41b2-be88-fa30eb0f4246" containerName="ovn-config" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.306595 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="21399e84-99fe-41b2-be88-fa30eb0f4246" containerName="ovn-config" Oct 08 09:27:57 crc kubenswrapper[4744]: E1008 09:27:57.306618 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f33787b3-d366-41b9-80f7-955dc458cb93" containerName="glance-db-sync" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.306626 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f33787b3-d366-41b9-80f7-955dc458cb93" containerName="glance-db-sync" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.306887 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="21399e84-99fe-41b2-be88-fa30eb0f4246" containerName="ovn-config" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.306912 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f33787b3-d366-41b9-80f7-955dc458cb93" containerName="glance-db-sync" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.308213 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.335632 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq"] Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.461131 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5mxg\" (UniqueName: \"kubernetes.io/projected/395ed1f9-be4b-482a-ba22-7623eecb44d6-kube-api-access-x5mxg\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.461433 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-ovsdbserver-sb\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.461960 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-dns-svc\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.462082 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-ovsdbserver-nb\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.462174 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-config\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.467495 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21399e84-99fe-41b2-be88-fa30eb0f4246" path="/var/lib/kubelet/pods/21399e84-99fe-41b2-be88-fa30eb0f4246/volumes" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.564843 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-ovsdbserver-nb\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.564940 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-config\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.565011 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5mxg\" (UniqueName: \"kubernetes.io/projected/395ed1f9-be4b-482a-ba22-7623eecb44d6-kube-api-access-x5mxg\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.565074 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-ovsdbserver-sb\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.565159 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-dns-svc\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.566668 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-dns-svc\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.567359 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-ovsdbserver-nb\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.568333 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-config\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.568922 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-ovsdbserver-sb\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.592986 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5mxg\" (UniqueName: \"kubernetes.io/projected/395ed1f9-be4b-482a-ba22-7623eecb44d6-kube-api-access-x5mxg\") pod \"dnsmasq-dns-6b5cb8c7cf-9gdbq\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:57 crc kubenswrapper[4744]: I1008 09:27:57.627353 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:58 crc kubenswrapper[4744]: I1008 09:27:58.195455 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq"] Oct 08 09:27:58 crc kubenswrapper[4744]: W1008 09:27:58.201629 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod395ed1f9_be4b_482a_ba22_7623eecb44d6.slice/crio-2f724d01479c2ef3e08ef186069e9b0339793c864a857ad317b3b17d3c3b73bb WatchSource:0}: Error finding container 2f724d01479c2ef3e08ef186069e9b0339793c864a857ad317b3b17d3c3b73bb: Status 404 returned error can't find the container with id 2f724d01479c2ef3e08ef186069e9b0339793c864a857ad317b3b17d3c3b73bb Oct 08 09:27:58 crc kubenswrapper[4744]: I1008 09:27:58.756159 4744 generic.go:334] "Generic (PLEG): container finished" podID="395ed1f9-be4b-482a-ba22-7623eecb44d6" containerID="4a732de79ee5e3345c73aa7ae53e7b04d65649de5902781c1a2937bf75137412" exitCode=0 Oct 08 09:27:58 crc kubenswrapper[4744]: I1008 09:27:58.756276 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" event={"ID":"395ed1f9-be4b-482a-ba22-7623eecb44d6","Type":"ContainerDied","Data":"4a732de79ee5e3345c73aa7ae53e7b04d65649de5902781c1a2937bf75137412"} Oct 08 09:27:58 crc kubenswrapper[4744]: I1008 09:27:58.756692 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" event={"ID":"395ed1f9-be4b-482a-ba22-7623eecb44d6","Type":"ContainerStarted","Data":"2f724d01479c2ef3e08ef186069e9b0339793c864a857ad317b3b17d3c3b73bb"} Oct 08 09:27:59 crc kubenswrapper[4744]: I1008 09:27:59.773619 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" event={"ID":"395ed1f9-be4b-482a-ba22-7623eecb44d6","Type":"ContainerStarted","Data":"0e72b9aa009c9fbe7c47cf736ce7c073904482761dba03f427c8bb92ab81380e"} Oct 08 09:27:59 crc kubenswrapper[4744]: I1008 09:27:59.775339 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:27:59 crc kubenswrapper[4744]: I1008 09:27:59.805489 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" podStartSLOduration=2.805467348 podStartE2EDuration="2.805467348s" podCreationTimestamp="2025-10-08 09:27:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:27:59.798447458 +0000 UTC m=+975.046092697" watchObservedRunningTime="2025-10-08 09:27:59.805467348 +0000 UTC m=+975.053112587" Oct 08 09:28:05 crc kubenswrapper[4744]: I1008 09:28:05.659587 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.003476 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-g7pft"] Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.004466 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-g7pft" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.021051 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-g7pft"] Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.028502 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsms5\" (UniqueName: \"kubernetes.io/projected/73553460-0850-41da-8ae1-33ced6d028f7-kube-api-access-jsms5\") pod \"barbican-db-create-g7pft\" (UID: \"73553460-0850-41da-8ae1-33ced6d028f7\") " pod="openstack/barbican-db-create-g7pft" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.070547 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.130114 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsms5\" (UniqueName: \"kubernetes.io/projected/73553460-0850-41da-8ae1-33ced6d028f7-kube-api-access-jsms5\") pod \"barbican-db-create-g7pft\" (UID: \"73553460-0850-41da-8ae1-33ced6d028f7\") " pod="openstack/barbican-db-create-g7pft" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.167172 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsms5\" (UniqueName: \"kubernetes.io/projected/73553460-0850-41da-8ae1-33ced6d028f7-kube-api-access-jsms5\") pod \"barbican-db-create-g7pft\" (UID: \"73553460-0850-41da-8ae1-33ced6d028f7\") " pod="openstack/barbican-db-create-g7pft" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.279830 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-pv2cr"] Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.281190 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pv2cr" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.295950 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-pv2cr"] Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.322678 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-g7pft" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.333486 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8mvn\" (UniqueName: \"kubernetes.io/projected/296ff709-3fab-4f25-ba36-7476d1a2b9d1-kube-api-access-s8mvn\") pod \"cinder-db-create-pv2cr\" (UID: \"296ff709-3fab-4f25-ba36-7476d1a2b9d1\") " pod="openstack/cinder-db-create-pv2cr" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.434689 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8mvn\" (UniqueName: \"kubernetes.io/projected/296ff709-3fab-4f25-ba36-7476d1a2b9d1-kube-api-access-s8mvn\") pod \"cinder-db-create-pv2cr\" (UID: \"296ff709-3fab-4f25-ba36-7476d1a2b9d1\") " pod="openstack/cinder-db-create-pv2cr" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.449244 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-7rjbs"] Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.450474 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7rjbs" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.460919 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-7rjbs"] Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.568412 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8mvn\" (UniqueName: \"kubernetes.io/projected/296ff709-3fab-4f25-ba36-7476d1a2b9d1-kube-api-access-s8mvn\") pod \"cinder-db-create-pv2cr\" (UID: \"296ff709-3fab-4f25-ba36-7476d1a2b9d1\") " pod="openstack/cinder-db-create-pv2cr" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.572770 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8tqw\" (UniqueName: \"kubernetes.io/projected/7706a783-6b39-431f-b765-e189122a38e2-kube-api-access-t8tqw\") pod \"neutron-db-create-7rjbs\" (UID: \"7706a783-6b39-431f-b765-e189122a38e2\") " pod="openstack/neutron-db-create-7rjbs" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.606084 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pv2cr" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.679262 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8tqw\" (UniqueName: \"kubernetes.io/projected/7706a783-6b39-431f-b765-e189122a38e2-kube-api-access-t8tqw\") pod \"neutron-db-create-7rjbs\" (UID: \"7706a783-6b39-431f-b765-e189122a38e2\") " pod="openstack/neutron-db-create-7rjbs" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.679837 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-zhx7v"] Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.680865 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.694143 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dkhdj" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.694422 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.700133 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.718606 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.727413 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zhx7v"] Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.742076 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8tqw\" (UniqueName: \"kubernetes.io/projected/7706a783-6b39-431f-b765-e189122a38e2-kube-api-access-t8tqw\") pod \"neutron-db-create-7rjbs\" (UID: \"7706a783-6b39-431f-b765-e189122a38e2\") " pod="openstack/neutron-db-create-7rjbs" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.783653 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-combined-ca-bundle\") pod \"keystone-db-sync-zhx7v\" (UID: \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\") " pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.783690 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-config-data\") pod \"keystone-db-sync-zhx7v\" (UID: \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\") " pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.783757 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwgfg\" (UniqueName: \"kubernetes.io/projected/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-kube-api-access-fwgfg\") pod \"keystone-db-sync-zhx7v\" (UID: \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\") " pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:06 crc kubenswrapper[4744]: I1008 09:28:06.870492 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7rjbs" Oct 08 09:28:07 crc kubenswrapper[4744]: I1008 09:28:06.884911 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwgfg\" (UniqueName: \"kubernetes.io/projected/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-kube-api-access-fwgfg\") pod \"keystone-db-sync-zhx7v\" (UID: \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\") " pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:07 crc kubenswrapper[4744]: I1008 09:28:06.884995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-combined-ca-bundle\") pod \"keystone-db-sync-zhx7v\" (UID: \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\") " pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:07 crc kubenswrapper[4744]: I1008 09:28:06.885018 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-config-data\") pod \"keystone-db-sync-zhx7v\" (UID: \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\") " pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:07 crc kubenswrapper[4744]: I1008 09:28:06.888601 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-config-data\") pod \"keystone-db-sync-zhx7v\" (UID: \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\") " pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:07 crc kubenswrapper[4744]: I1008 09:28:06.894231 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-combined-ca-bundle\") pod \"keystone-db-sync-zhx7v\" (UID: \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\") " pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:07 crc kubenswrapper[4744]: I1008 09:28:06.902518 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwgfg\" (UniqueName: \"kubernetes.io/projected/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-kube-api-access-fwgfg\") pod \"keystone-db-sync-zhx7v\" (UID: \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\") " pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:07 crc kubenswrapper[4744]: I1008 09:28:07.024089 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:07 crc kubenswrapper[4744]: I1008 09:28:07.629585 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:28:07 crc kubenswrapper[4744]: I1008 09:28:07.690911 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-lvxbs"] Oct 08 09:28:07 crc kubenswrapper[4744]: I1008 09:28:07.691358 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" podUID="45d914c8-25de-4c20-84db-2d26c95289ee" containerName="dnsmasq-dns" containerID="cri-o://9b2405648a5422cb7d36175a47c300fe650aeb02ee481e5cd9de79b6929756eb" gracePeriod=10 Oct 08 09:28:07 crc kubenswrapper[4744]: I1008 09:28:07.880578 4744 generic.go:334] "Generic (PLEG): container finished" podID="45d914c8-25de-4c20-84db-2d26c95289ee" containerID="9b2405648a5422cb7d36175a47c300fe650aeb02ee481e5cd9de79b6929756eb" exitCode=0 Oct 08 09:28:07 crc kubenswrapper[4744]: I1008 09:28:07.880621 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" event={"ID":"45d914c8-25de-4c20-84db-2d26c95289ee","Type":"ContainerDied","Data":"9b2405648a5422cb7d36175a47c300fe650aeb02ee481e5cd9de79b6929756eb"} Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.032691 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-g7pft"] Oct 08 09:28:08 crc kubenswrapper[4744]: W1008 09:28:08.359186 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7706a783_6b39_431f_b765_e189122a38e2.slice/crio-27a37519ad68442fc4e2c32f4e1926939eff1267fb6ac4485756416aa17d4861 WatchSource:0}: Error finding container 27a37519ad68442fc4e2c32f4e1926939eff1267fb6ac4485756416aa17d4861: Status 404 returned error can't find the container with id 27a37519ad68442fc4e2c32f4e1926939eff1267fb6ac4485756416aa17d4861 Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.362579 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-7rjbs"] Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.372553 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-pv2cr"] Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.393204 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zhx7v"] Oct 08 09:28:08 crc kubenswrapper[4744]: W1008 09:28:08.405243 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod296ff709_3fab_4f25_ba36_7476d1a2b9d1.slice/crio-38e523999e9a5dd52c3a1c96f62247aab48c9478561a0f93edf2a8c71a43dd53 WatchSource:0}: Error finding container 38e523999e9a5dd52c3a1c96f62247aab48c9478561a0f93edf2a8c71a43dd53: Status 404 returned error can't find the container with id 38e523999e9a5dd52c3a1c96f62247aab48c9478561a0f93edf2a8c71a43dd53 Oct 08 09:28:08 crc kubenswrapper[4744]: W1008 09:28:08.412764 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbab2aef1_a444_4b40_bcc3_b59ba4755bf7.slice/crio-08e7bb30eef4e632690310fdbef95e944a7650b6b14691673d6242eaa2b2b02f WatchSource:0}: Error finding container 08e7bb30eef4e632690310fdbef95e944a7650b6b14691673d6242eaa2b2b02f: Status 404 returned error can't find the container with id 08e7bb30eef4e632690310fdbef95e944a7650b6b14691673d6242eaa2b2b02f Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.584771 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.621770 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-config\") pod \"45d914c8-25de-4c20-84db-2d26c95289ee\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.622793 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-ovsdbserver-nb\") pod \"45d914c8-25de-4c20-84db-2d26c95289ee\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.622871 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-ovsdbserver-sb\") pod \"45d914c8-25de-4c20-84db-2d26c95289ee\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.622943 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-dns-svc\") pod \"45d914c8-25de-4c20-84db-2d26c95289ee\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.623061 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlwl2\" (UniqueName: \"kubernetes.io/projected/45d914c8-25de-4c20-84db-2d26c95289ee-kube-api-access-hlwl2\") pod \"45d914c8-25de-4c20-84db-2d26c95289ee\" (UID: \"45d914c8-25de-4c20-84db-2d26c95289ee\") " Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.638625 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45d914c8-25de-4c20-84db-2d26c95289ee-kube-api-access-hlwl2" (OuterVolumeSpecName: "kube-api-access-hlwl2") pod "45d914c8-25de-4c20-84db-2d26c95289ee" (UID: "45d914c8-25de-4c20-84db-2d26c95289ee"). InnerVolumeSpecName "kube-api-access-hlwl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.724035 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlwl2\" (UniqueName: \"kubernetes.io/projected/45d914c8-25de-4c20-84db-2d26c95289ee-kube-api-access-hlwl2\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.763553 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-config" (OuterVolumeSpecName: "config") pod "45d914c8-25de-4c20-84db-2d26c95289ee" (UID: "45d914c8-25de-4c20-84db-2d26c95289ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.786550 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "45d914c8-25de-4c20-84db-2d26c95289ee" (UID: "45d914c8-25de-4c20-84db-2d26c95289ee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.792998 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "45d914c8-25de-4c20-84db-2d26c95289ee" (UID: "45d914c8-25de-4c20-84db-2d26c95289ee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.801578 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "45d914c8-25de-4c20-84db-2d26c95289ee" (UID: "45d914c8-25de-4c20-84db-2d26c95289ee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.825163 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.825206 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.825220 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.825231 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/45d914c8-25de-4c20-84db-2d26c95289ee-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.893876 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zhx7v" event={"ID":"bab2aef1-a444-4b40-bcc3-b59ba4755bf7","Type":"ContainerStarted","Data":"08e7bb30eef4e632690310fdbef95e944a7650b6b14691673d6242eaa2b2b02f"} Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.895719 4744 generic.go:334] "Generic (PLEG): container finished" podID="296ff709-3fab-4f25-ba36-7476d1a2b9d1" containerID="c7a6b503518b670ce3301e8c0a603a1738d812039d0262a6307ad03d6cbbab63" exitCode=0 Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.895768 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pv2cr" event={"ID":"296ff709-3fab-4f25-ba36-7476d1a2b9d1","Type":"ContainerDied","Data":"c7a6b503518b670ce3301e8c0a603a1738d812039d0262a6307ad03d6cbbab63"} Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.895785 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pv2cr" event={"ID":"296ff709-3fab-4f25-ba36-7476d1a2b9d1","Type":"ContainerStarted","Data":"38e523999e9a5dd52c3a1c96f62247aab48c9478561a0f93edf2a8c71a43dd53"} Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.898280 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" event={"ID":"45d914c8-25de-4c20-84db-2d26c95289ee","Type":"ContainerDied","Data":"e3790c372ceb3218d3b98287b21e9a25458e5f97d92deab71ec6b94e505a0aca"} Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.898323 4744 scope.go:117] "RemoveContainer" containerID="9b2405648a5422cb7d36175a47c300fe650aeb02ee481e5cd9de79b6929756eb" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.898498 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc45f6dcf-lvxbs" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.900730 4744 generic.go:334] "Generic (PLEG): container finished" podID="73553460-0850-41da-8ae1-33ced6d028f7" containerID="b2a66635ddeb7399d57dd21ebf682ba5b1ca75fc8934a7145df8d3f597cffb92" exitCode=0 Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.900806 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-g7pft" event={"ID":"73553460-0850-41da-8ae1-33ced6d028f7","Type":"ContainerDied","Data":"b2a66635ddeb7399d57dd21ebf682ba5b1ca75fc8934a7145df8d3f597cffb92"} Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.900828 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-g7pft" event={"ID":"73553460-0850-41da-8ae1-33ced6d028f7","Type":"ContainerStarted","Data":"4fb422260cc3a39fb0147dca96fb0e7124877e1c940f4085999c189b474395c8"} Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.902822 4744 generic.go:334] "Generic (PLEG): container finished" podID="7706a783-6b39-431f-b765-e189122a38e2" containerID="5e8fb6af05a95d7d62e576edbfb946d591709c6c28b7dd46787bdf59e83da84d" exitCode=0 Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.902945 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7rjbs" event={"ID":"7706a783-6b39-431f-b765-e189122a38e2","Type":"ContainerDied","Data":"5e8fb6af05a95d7d62e576edbfb946d591709c6c28b7dd46787bdf59e83da84d"} Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.902966 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7rjbs" event={"ID":"7706a783-6b39-431f-b765-e189122a38e2","Type":"ContainerStarted","Data":"27a37519ad68442fc4e2c32f4e1926939eff1267fb6ac4485756416aa17d4861"} Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.921253 4744 scope.go:117] "RemoveContainer" containerID="89ef0bc8c22be55fcfb810002c146fc9173bbec232942305725d3d8739ef3622" Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.974624 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-lvxbs"] Oct 08 09:28:08 crc kubenswrapper[4744]: I1008 09:28:08.981080 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bc45f6dcf-lvxbs"] Oct 08 09:28:09 crc kubenswrapper[4744]: I1008 09:28:09.462044 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45d914c8-25de-4c20-84db-2d26c95289ee" path="/var/lib/kubelet/pods/45d914c8-25de-4c20-84db-2d26c95289ee/volumes" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.305193 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7rjbs" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.450695 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-g7pft" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.462248 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8tqw\" (UniqueName: \"kubernetes.io/projected/7706a783-6b39-431f-b765-e189122a38e2-kube-api-access-t8tqw\") pod \"7706a783-6b39-431f-b765-e189122a38e2\" (UID: \"7706a783-6b39-431f-b765-e189122a38e2\") " Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.476323 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7706a783-6b39-431f-b765-e189122a38e2-kube-api-access-t8tqw" (OuterVolumeSpecName: "kube-api-access-t8tqw") pod "7706a783-6b39-431f-b765-e189122a38e2" (UID: "7706a783-6b39-431f-b765-e189122a38e2"). InnerVolumeSpecName "kube-api-access-t8tqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.534147 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pv2cr" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.564609 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8mvn\" (UniqueName: \"kubernetes.io/projected/296ff709-3fab-4f25-ba36-7476d1a2b9d1-kube-api-access-s8mvn\") pod \"296ff709-3fab-4f25-ba36-7476d1a2b9d1\" (UID: \"296ff709-3fab-4f25-ba36-7476d1a2b9d1\") " Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.564654 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsms5\" (UniqueName: \"kubernetes.io/projected/73553460-0850-41da-8ae1-33ced6d028f7-kube-api-access-jsms5\") pod \"73553460-0850-41da-8ae1-33ced6d028f7\" (UID: \"73553460-0850-41da-8ae1-33ced6d028f7\") " Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.564929 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8tqw\" (UniqueName: \"kubernetes.io/projected/7706a783-6b39-431f-b765-e189122a38e2-kube-api-access-t8tqw\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.568331 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73553460-0850-41da-8ae1-33ced6d028f7-kube-api-access-jsms5" (OuterVolumeSpecName: "kube-api-access-jsms5") pod "73553460-0850-41da-8ae1-33ced6d028f7" (UID: "73553460-0850-41da-8ae1-33ced6d028f7"). InnerVolumeSpecName "kube-api-access-jsms5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.583435 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/296ff709-3fab-4f25-ba36-7476d1a2b9d1-kube-api-access-s8mvn" (OuterVolumeSpecName: "kube-api-access-s8mvn") pod "296ff709-3fab-4f25-ba36-7476d1a2b9d1" (UID: "296ff709-3fab-4f25-ba36-7476d1a2b9d1"). InnerVolumeSpecName "kube-api-access-s8mvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.666760 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8mvn\" (UniqueName: \"kubernetes.io/projected/296ff709-3fab-4f25-ba36-7476d1a2b9d1-kube-api-access-s8mvn\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.666802 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsms5\" (UniqueName: \"kubernetes.io/projected/73553460-0850-41da-8ae1-33ced6d028f7-kube-api-access-jsms5\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.923799 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-g7pft" event={"ID":"73553460-0850-41da-8ae1-33ced6d028f7","Type":"ContainerDied","Data":"4fb422260cc3a39fb0147dca96fb0e7124877e1c940f4085999c189b474395c8"} Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.924122 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fb422260cc3a39fb0147dca96fb0e7124877e1c940f4085999c189b474395c8" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.924187 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-g7pft" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.934787 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-7rjbs" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.934826 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-7rjbs" event={"ID":"7706a783-6b39-431f-b765-e189122a38e2","Type":"ContainerDied","Data":"27a37519ad68442fc4e2c32f4e1926939eff1267fb6ac4485756416aa17d4861"} Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.935089 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27a37519ad68442fc4e2c32f4e1926939eff1267fb6ac4485756416aa17d4861" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.939966 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pv2cr" event={"ID":"296ff709-3fab-4f25-ba36-7476d1a2b9d1","Type":"ContainerDied","Data":"38e523999e9a5dd52c3a1c96f62247aab48c9478561a0f93edf2a8c71a43dd53"} Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.940006 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38e523999e9a5dd52c3a1c96f62247aab48c9478561a0f93edf2a8c71a43dd53" Oct 08 09:28:10 crc kubenswrapper[4744]: I1008 09:28:10.940069 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pv2cr" Oct 08 09:28:14 crc kubenswrapper[4744]: I1008 09:28:14.995499 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zhx7v" event={"ID":"bab2aef1-a444-4b40-bcc3-b59ba4755bf7","Type":"ContainerStarted","Data":"b5e7b98c0e146c6e932e8602f8fcfd2d7d0c377be903294c753a386d73e111fb"} Oct 08 09:28:15 crc kubenswrapper[4744]: I1008 09:28:15.029451 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-zhx7v" podStartSLOduration=3.355410202 podStartE2EDuration="9.029422474s" podCreationTimestamp="2025-10-08 09:28:06 +0000 UTC" firstStartedPulling="2025-10-08 09:28:08.416386171 +0000 UTC m=+983.664031410" lastFinishedPulling="2025-10-08 09:28:14.090398433 +0000 UTC m=+989.338043682" observedRunningTime="2025-10-08 09:28:15.018356398 +0000 UTC m=+990.266001657" watchObservedRunningTime="2025-10-08 09:28:15.029422474 +0000 UTC m=+990.277067723" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.449118 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c3ca-account-create-nftjn"] Oct 08 09:28:16 crc kubenswrapper[4744]: E1008 09:28:16.451524 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="296ff709-3fab-4f25-ba36-7476d1a2b9d1" containerName="mariadb-database-create" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.451555 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="296ff709-3fab-4f25-ba36-7476d1a2b9d1" containerName="mariadb-database-create" Oct 08 09:28:16 crc kubenswrapper[4744]: E1008 09:28:16.451587 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7706a783-6b39-431f-b765-e189122a38e2" containerName="mariadb-database-create" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.451597 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7706a783-6b39-431f-b765-e189122a38e2" containerName="mariadb-database-create" Oct 08 09:28:16 crc kubenswrapper[4744]: E1008 09:28:16.451610 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d914c8-25de-4c20-84db-2d26c95289ee" containerName="dnsmasq-dns" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.451617 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d914c8-25de-4c20-84db-2d26c95289ee" containerName="dnsmasq-dns" Oct 08 09:28:16 crc kubenswrapper[4744]: E1008 09:28:16.451638 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73553460-0850-41da-8ae1-33ced6d028f7" containerName="mariadb-database-create" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.451650 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="73553460-0850-41da-8ae1-33ced6d028f7" containerName="mariadb-database-create" Oct 08 09:28:16 crc kubenswrapper[4744]: E1008 09:28:16.451667 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d914c8-25de-4c20-84db-2d26c95289ee" containerName="init" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.451675 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d914c8-25de-4c20-84db-2d26c95289ee" containerName="init" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.451885 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="296ff709-3fab-4f25-ba36-7476d1a2b9d1" containerName="mariadb-database-create" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.451913 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="73553460-0850-41da-8ae1-33ced6d028f7" containerName="mariadb-database-create" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.451927 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7706a783-6b39-431f-b765-e189122a38e2" containerName="mariadb-database-create" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.451948 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d914c8-25de-4c20-84db-2d26c95289ee" containerName="dnsmasq-dns" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.455181 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c3ca-account-create-nftjn" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.463695 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.470523 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c3ca-account-create-nftjn"] Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.594175 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kh9h\" (UniqueName: \"kubernetes.io/projected/a658cb27-4e1d-4d3b-90cd-1e16b4013c48-kube-api-access-2kh9h\") pod \"neutron-c3ca-account-create-nftjn\" (UID: \"a658cb27-4e1d-4d3b-90cd-1e16b4013c48\") " pod="openstack/neutron-c3ca-account-create-nftjn" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.695124 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kh9h\" (UniqueName: \"kubernetes.io/projected/a658cb27-4e1d-4d3b-90cd-1e16b4013c48-kube-api-access-2kh9h\") pod \"neutron-c3ca-account-create-nftjn\" (UID: \"a658cb27-4e1d-4d3b-90cd-1e16b4013c48\") " pod="openstack/neutron-c3ca-account-create-nftjn" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.726228 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kh9h\" (UniqueName: \"kubernetes.io/projected/a658cb27-4e1d-4d3b-90cd-1e16b4013c48-kube-api-access-2kh9h\") pod \"neutron-c3ca-account-create-nftjn\" (UID: \"a658cb27-4e1d-4d3b-90cd-1e16b4013c48\") " pod="openstack/neutron-c3ca-account-create-nftjn" Oct 08 09:28:16 crc kubenswrapper[4744]: I1008 09:28:16.779084 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c3ca-account-create-nftjn" Oct 08 09:28:17 crc kubenswrapper[4744]: I1008 09:28:17.286837 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c3ca-account-create-nftjn"] Oct 08 09:28:18 crc kubenswrapper[4744]: I1008 09:28:18.031214 4744 generic.go:334] "Generic (PLEG): container finished" podID="a658cb27-4e1d-4d3b-90cd-1e16b4013c48" containerID="4e736fd721b1c7fc0f95275c73e3d9f42b2f078ba886f2fc7f4bc07720e721a4" exitCode=0 Oct 08 09:28:18 crc kubenswrapper[4744]: I1008 09:28:18.031309 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c3ca-account-create-nftjn" event={"ID":"a658cb27-4e1d-4d3b-90cd-1e16b4013c48","Type":"ContainerDied","Data":"4e736fd721b1c7fc0f95275c73e3d9f42b2f078ba886f2fc7f4bc07720e721a4"} Oct 08 09:28:18 crc kubenswrapper[4744]: I1008 09:28:18.031697 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c3ca-account-create-nftjn" event={"ID":"a658cb27-4e1d-4d3b-90cd-1e16b4013c48","Type":"ContainerStarted","Data":"e791698aa72df212def3737460ab682b778e1637b7e4fc10bf2c55e717833bf6"} Oct 08 09:28:19 crc kubenswrapper[4744]: I1008 09:28:19.046582 4744 generic.go:334] "Generic (PLEG): container finished" podID="bab2aef1-a444-4b40-bcc3-b59ba4755bf7" containerID="b5e7b98c0e146c6e932e8602f8fcfd2d7d0c377be903294c753a386d73e111fb" exitCode=0 Oct 08 09:28:19 crc kubenswrapper[4744]: I1008 09:28:19.046662 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zhx7v" event={"ID":"bab2aef1-a444-4b40-bcc3-b59ba4755bf7","Type":"ContainerDied","Data":"b5e7b98c0e146c6e932e8602f8fcfd2d7d0c377be903294c753a386d73e111fb"} Oct 08 09:28:19 crc kubenswrapper[4744]: I1008 09:28:19.439114 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c3ca-account-create-nftjn" Oct 08 09:28:19 crc kubenswrapper[4744]: I1008 09:28:19.556314 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kh9h\" (UniqueName: \"kubernetes.io/projected/a658cb27-4e1d-4d3b-90cd-1e16b4013c48-kube-api-access-2kh9h\") pod \"a658cb27-4e1d-4d3b-90cd-1e16b4013c48\" (UID: \"a658cb27-4e1d-4d3b-90cd-1e16b4013c48\") " Oct 08 09:28:19 crc kubenswrapper[4744]: I1008 09:28:19.569613 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a658cb27-4e1d-4d3b-90cd-1e16b4013c48-kube-api-access-2kh9h" (OuterVolumeSpecName: "kube-api-access-2kh9h") pod "a658cb27-4e1d-4d3b-90cd-1e16b4013c48" (UID: "a658cb27-4e1d-4d3b-90cd-1e16b4013c48"). InnerVolumeSpecName "kube-api-access-2kh9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:19 crc kubenswrapper[4744]: I1008 09:28:19.658998 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kh9h\" (UniqueName: \"kubernetes.io/projected/a658cb27-4e1d-4d3b-90cd-1e16b4013c48-kube-api-access-2kh9h\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.065257 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c3ca-account-create-nftjn" event={"ID":"a658cb27-4e1d-4d3b-90cd-1e16b4013c48","Type":"ContainerDied","Data":"e791698aa72df212def3737460ab682b778e1637b7e4fc10bf2c55e717833bf6"} Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.065324 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e791698aa72df212def3737460ab682b778e1637b7e4fc10bf2c55e717833bf6" Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.065331 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c3ca-account-create-nftjn" Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.480944 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.575262 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-combined-ca-bundle\") pod \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\" (UID: \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\") " Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.575544 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwgfg\" (UniqueName: \"kubernetes.io/projected/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-kube-api-access-fwgfg\") pod \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\" (UID: \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\") " Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.575685 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-config-data\") pod \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\" (UID: \"bab2aef1-a444-4b40-bcc3-b59ba4755bf7\") " Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.594984 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-kube-api-access-fwgfg" (OuterVolumeSpecName: "kube-api-access-fwgfg") pod "bab2aef1-a444-4b40-bcc3-b59ba4755bf7" (UID: "bab2aef1-a444-4b40-bcc3-b59ba4755bf7"). InnerVolumeSpecName "kube-api-access-fwgfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.608344 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bab2aef1-a444-4b40-bcc3-b59ba4755bf7" (UID: "bab2aef1-a444-4b40-bcc3-b59ba4755bf7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.626503 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-config-data" (OuterVolumeSpecName: "config-data") pod "bab2aef1-a444-4b40-bcc3-b59ba4755bf7" (UID: "bab2aef1-a444-4b40-bcc3-b59ba4755bf7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.679133 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.679178 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:20 crc kubenswrapper[4744]: I1008 09:28:20.679192 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwgfg\" (UniqueName: \"kubernetes.io/projected/bab2aef1-a444-4b40-bcc3-b59ba4755bf7-kube-api-access-fwgfg\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.077240 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zhx7v" event={"ID":"bab2aef1-a444-4b40-bcc3-b59ba4755bf7","Type":"ContainerDied","Data":"08e7bb30eef4e632690310fdbef95e944a7650b6b14691673d6242eaa2b2b02f"} Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.077773 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08e7bb30eef4e632690310fdbef95e944a7650b6b14691673d6242eaa2b2b02f" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.078023 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zhx7v" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.389822 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6p2wt"] Oct 08 09:28:21 crc kubenswrapper[4744]: E1008 09:28:21.390113 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a658cb27-4e1d-4d3b-90cd-1e16b4013c48" containerName="mariadb-account-create" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.390126 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a658cb27-4e1d-4d3b-90cd-1e16b4013c48" containerName="mariadb-account-create" Oct 08 09:28:21 crc kubenswrapper[4744]: E1008 09:28:21.390151 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bab2aef1-a444-4b40-bcc3-b59ba4755bf7" containerName="keystone-db-sync" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.390156 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bab2aef1-a444-4b40-bcc3-b59ba4755bf7" containerName="keystone-db-sync" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.390315 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bab2aef1-a444-4b40-bcc3-b59ba4755bf7" containerName="keystone-db-sync" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.390332 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a658cb27-4e1d-4d3b-90cd-1e16b4013c48" containerName="mariadb-account-create" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.390818 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.395210 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.395411 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.399077 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.399230 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dkhdj" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.445887 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c7d6ff65f-28q4g"] Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.502249 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-scripts\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.502402 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-fernet-keys\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.502543 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mjqt\" (UniqueName: \"kubernetes.io/projected/bb95350d-9809-47d8-9fd7-ae44753efe2f-kube-api-access-6mjqt\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.502625 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-credential-keys\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.502650 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-combined-ca-bundle\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.502711 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-config-data\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.577665 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.602833 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6p2wt"] Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.608124 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-fernet-keys\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.608224 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mjqt\" (UniqueName: \"kubernetes.io/projected/bb95350d-9809-47d8-9fd7-ae44753efe2f-kube-api-access-6mjqt\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.608307 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-credential-keys\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.608330 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-combined-ca-bundle\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.608426 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-config-data\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.608549 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-scripts\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.617208 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c7d6ff65f-28q4g"] Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.618103 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-fernet-keys\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.623144 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-scripts\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.623648 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-combined-ca-bundle\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.635101 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-credential-keys\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.637677 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-config-data\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.642983 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mjqt\" (UniqueName: \"kubernetes.io/projected/bb95350d-9809-47d8-9fd7-ae44753efe2f-kube-api-access-6mjqt\") pod \"keystone-bootstrap-6p2wt\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.709639 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-dns-svc\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.709853 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-ovsdbserver-nb\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.709956 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-config\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.710073 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-ovsdbserver-sb\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.710299 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg2ls\" (UniqueName: \"kubernetes.io/projected/9563167e-7b47-4b29-85e8-4e299c9e10d3-kube-api-access-jg2ls\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.720025 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.807937 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.812052 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-dns-svc\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.812215 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-ovsdbserver-nb\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.812249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-config\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.812320 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-ovsdbserver-sb\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.812414 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg2ls\" (UniqueName: \"kubernetes.io/projected/9563167e-7b47-4b29-85e8-4e299c9e10d3-kube-api-access-jg2ls\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.814337 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-dns-svc\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.822571 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-ovsdbserver-nb\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.825530 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-config\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.833738 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.840715 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-ovsdbserver-sb\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.842297 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.846705 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.869993 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg2ls\" (UniqueName: \"kubernetes.io/projected/9563167e-7b47-4b29-85e8-4e299c9e10d3-kube-api-access-jg2ls\") pod \"dnsmasq-dns-c7d6ff65f-28q4g\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.887195 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.906924 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.922634 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-scripts\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.922766 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-config-data\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.922832 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04474f89-9bb4-4bb7-aef3-66d519650a4b-log-httpd\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.922931 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrjgk\" (UniqueName: \"kubernetes.io/projected/04474f89-9bb4-4bb7-aef3-66d519650a4b-kube-api-access-vrjgk\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.923039 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.923080 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.923124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04474f89-9bb4-4bb7-aef3-66d519650a4b-run-httpd\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.987490 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-fxgnx"] Oct 08 09:28:21 crc kubenswrapper[4744]: I1008 09:28:21.989231 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.015091 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-bp6tb" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.015530 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.023848 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.025199 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04474f89-9bb4-4bb7-aef3-66d519650a4b-log-httpd\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.025261 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrjgk\" (UniqueName: \"kubernetes.io/projected/04474f89-9bb4-4bb7-aef3-66d519650a4b-kube-api-access-vrjgk\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.025305 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.025328 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.025348 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04474f89-9bb4-4bb7-aef3-66d519650a4b-run-httpd\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.025406 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-scripts\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.025429 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-config-data\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.036201 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04474f89-9bb4-4bb7-aef3-66d519650a4b-run-httpd\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.058172 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04474f89-9bb4-4bb7-aef3-66d519650a4b-log-httpd\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.061053 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.063279 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-config-data\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.066513 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.119005 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-scripts\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.121188 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrjgk\" (UniqueName: \"kubernetes.io/projected/04474f89-9bb4-4bb7-aef3-66d519650a4b-kube-api-access-vrjgk\") pod \"ceilometer-0\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.127489 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-fxgnx"] Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.137343 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hth5h\" (UniqueName: \"kubernetes.io/projected/54a05787-2fe8-4d6d-b785-058040260352-kube-api-access-hth5h\") pod \"neutron-db-sync-fxgnx\" (UID: \"54a05787-2fe8-4d6d-b785-058040260352\") " pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.137455 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a05787-2fe8-4d6d-b785-058040260352-combined-ca-bundle\") pod \"neutron-db-sync-fxgnx\" (UID: \"54a05787-2fe8-4d6d-b785-058040260352\") " pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.137475 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/54a05787-2fe8-4d6d-b785-058040260352-config\") pod \"neutron-db-sync-fxgnx\" (UID: \"54a05787-2fe8-4d6d-b785-058040260352\") " pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.206101 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-g7tp7"] Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.211442 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.223213 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.225129 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.225788 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-89lxn" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.230863 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.239326 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a05787-2fe8-4d6d-b785-058040260352-combined-ca-bundle\") pod \"neutron-db-sync-fxgnx\" (UID: \"54a05787-2fe8-4d6d-b785-058040260352\") " pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.239404 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/54a05787-2fe8-4d6d-b785-058040260352-config\") pod \"neutron-db-sync-fxgnx\" (UID: \"54a05787-2fe8-4d6d-b785-058040260352\") " pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.239569 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hth5h\" (UniqueName: \"kubernetes.io/projected/54a05787-2fe8-4d6d-b785-058040260352-kube-api-access-hth5h\") pod \"neutron-db-sync-fxgnx\" (UID: \"54a05787-2fe8-4d6d-b785-058040260352\") " pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.247565 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c7d6ff65f-28q4g"] Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.252180 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a05787-2fe8-4d6d-b785-058040260352-combined-ca-bundle\") pod \"neutron-db-sync-fxgnx\" (UID: \"54a05787-2fe8-4d6d-b785-058040260352\") " pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.265111 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/54a05787-2fe8-4d6d-b785-058040260352-config\") pod \"neutron-db-sync-fxgnx\" (UID: \"54a05787-2fe8-4d6d-b785-058040260352\") " pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.285955 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-g7tp7"] Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.316819 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hth5h\" (UniqueName: \"kubernetes.io/projected/54a05787-2fe8-4d6d-b785-058040260352-kube-api-access-hth5h\") pod \"neutron-db-sync-fxgnx\" (UID: \"54a05787-2fe8-4d6d-b785-058040260352\") " pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.342824 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-748d7644cf-c74hj"] Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.343171 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-combined-ca-bundle\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.343232 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6a227c0-ae15-427c-9557-72be160cb20c-logs\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.343273 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-config-data\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.343297 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6gsz\" (UniqueName: \"kubernetes.io/projected/a6a227c0-ae15-427c-9557-72be160cb20c-kube-api-access-r6gsz\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.343323 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-scripts\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.350384 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.393033 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.413503 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-748d7644cf-c74hj"] Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.450216 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-config-data\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.450287 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6gsz\" (UniqueName: \"kubernetes.io/projected/a6a227c0-ae15-427c-9557-72be160cb20c-kube-api-access-r6gsz\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.450314 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-config\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.450337 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-scripts\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.450390 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-ovsdbserver-sb\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.450441 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4zlk\" (UniqueName: \"kubernetes.io/projected/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-kube-api-access-p4zlk\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.450511 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-ovsdbserver-nb\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.450549 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-combined-ca-bundle\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.450571 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6a227c0-ae15-427c-9557-72be160cb20c-logs\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.450600 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-dns-svc\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.471643 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6a227c0-ae15-427c-9557-72be160cb20c-logs\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.488003 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6gsz\" (UniqueName: \"kubernetes.io/projected/a6a227c0-ae15-427c-9557-72be160cb20c-kube-api-access-r6gsz\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.488467 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-config-data\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.489389 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-scripts\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.496441 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-combined-ca-bundle\") pod \"placement-db-sync-g7tp7\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.557213 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-config\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.557295 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-ovsdbserver-sb\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.557349 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4zlk\" (UniqueName: \"kubernetes.io/projected/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-kube-api-access-p4zlk\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.557432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-ovsdbserver-nb\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.557511 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-dns-svc\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.558525 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-dns-svc\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.559252 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-ovsdbserver-nb\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.559261 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-config\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.559842 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-ovsdbserver-sb\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.577715 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.602149 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4zlk\" (UniqueName: \"kubernetes.io/projected/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-kube-api-access-p4zlk\") pod \"dnsmasq-dns-748d7644cf-c74hj\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.636952 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6p2wt"] Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.698393 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:22 crc kubenswrapper[4744]: I1008 09:28:22.936261 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c7d6ff65f-28q4g"] Oct 08 09:28:23 crc kubenswrapper[4744]: I1008 09:28:23.182292 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" event={"ID":"9563167e-7b47-4b29-85e8-4e299c9e10d3","Type":"ContainerStarted","Data":"94e8c994ab7c5a786d19d9d831f1ade5e26699e43e857001f5233a55ff14e3b0"} Oct 08 09:28:23 crc kubenswrapper[4744]: I1008 09:28:23.189425 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6p2wt" event={"ID":"bb95350d-9809-47d8-9fd7-ae44753efe2f","Type":"ContainerStarted","Data":"aecd452ff4b9b42a52af9b1cf664ab56f7b6c6a088294b1d799d6af744fd3350"} Oct 08 09:28:23 crc kubenswrapper[4744]: I1008 09:28:23.385963 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:28:23 crc kubenswrapper[4744]: I1008 09:28:23.490463 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-fxgnx"] Oct 08 09:28:23 crc kubenswrapper[4744]: I1008 09:28:23.494349 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-g7tp7"] Oct 08 09:28:23 crc kubenswrapper[4744]: I1008 09:28:23.630735 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-748d7644cf-c74hj"] Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.206185 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fxgnx" event={"ID":"54a05787-2fe8-4d6d-b785-058040260352","Type":"ContainerStarted","Data":"2544b2efd54a6b3b29f077a1fd78405bc7739e74cf3844435e05fc27b230527e"} Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.206569 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fxgnx" event={"ID":"54a05787-2fe8-4d6d-b785-058040260352","Type":"ContainerStarted","Data":"35bd34f0c504bf49b11d3634c24956712872c125a0c7b06b63afd7201a305031"} Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.212528 4744 generic.go:334] "Generic (PLEG): container finished" podID="993f71a0-3367-4a19-b0b0-b7c75b4d7e16" containerID="8176f4ede4b55fe74a47acd764af270597934eb1d8deb1df71deabc9f9175ada" exitCode=0 Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.212617 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" event={"ID":"993f71a0-3367-4a19-b0b0-b7c75b4d7e16","Type":"ContainerDied","Data":"8176f4ede4b55fe74a47acd764af270597934eb1d8deb1df71deabc9f9175ada"} Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.212665 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" event={"ID":"993f71a0-3367-4a19-b0b0-b7c75b4d7e16","Type":"ContainerStarted","Data":"28b9fff0f3e701e110d9460cf9192b01141d52f5c9d9ea58a6e47bf8ee139de0"} Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.214445 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04474f89-9bb4-4bb7-aef3-66d519650a4b","Type":"ContainerStarted","Data":"8966f601de44cde4dbc5b42e98a3693dd7d4c8a7a817c295ae138720de185af4"} Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.216232 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-g7tp7" event={"ID":"a6a227c0-ae15-427c-9557-72be160cb20c","Type":"ContainerStarted","Data":"07b1b938882f375d48c1d59963a929e7d2e1d8b53df1dd18de562f8e05e5f055"} Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.218416 4744 generic.go:334] "Generic (PLEG): container finished" podID="9563167e-7b47-4b29-85e8-4e299c9e10d3" containerID="5ddfbda7ed0fecbd0c8f4e4f2496a77efbda95b751f843911a4e04da7a3c532c" exitCode=0 Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.218484 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" event={"ID":"9563167e-7b47-4b29-85e8-4e299c9e10d3","Type":"ContainerDied","Data":"5ddfbda7ed0fecbd0c8f4e4f2496a77efbda95b751f843911a4e04da7a3c532c"} Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.219531 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6p2wt" event={"ID":"bb95350d-9809-47d8-9fd7-ae44753efe2f","Type":"ContainerStarted","Data":"798489ddd7442f1d55d1aea17083ea9445108c87809654cc94d54563aa808f0c"} Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.251211 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-fxgnx" podStartSLOduration=3.251191812 podStartE2EDuration="3.251191812s" podCreationTimestamp="2025-10-08 09:28:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:28:24.228632589 +0000 UTC m=+999.476277838" watchObservedRunningTime="2025-10-08 09:28:24.251191812 +0000 UTC m=+999.498837041" Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.266719 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6p2wt" podStartSLOduration=3.266668124 podStartE2EDuration="3.266668124s" podCreationTimestamp="2025-10-08 09:28:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:28:24.25953214 +0000 UTC m=+999.507177379" watchObservedRunningTime="2025-10-08 09:28:24.266668124 +0000 UTC m=+999.514313363" Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.739660 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.813848 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.822554 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-ovsdbserver-sb\") pod \"9563167e-7b47-4b29-85e8-4e299c9e10d3\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.822649 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-ovsdbserver-nb\") pod \"9563167e-7b47-4b29-85e8-4e299c9e10d3\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.822696 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg2ls\" (UniqueName: \"kubernetes.io/projected/9563167e-7b47-4b29-85e8-4e299c9e10d3-kube-api-access-jg2ls\") pod \"9563167e-7b47-4b29-85e8-4e299c9e10d3\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.822722 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-dns-svc\") pod \"9563167e-7b47-4b29-85e8-4e299c9e10d3\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.822778 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-config\") pod \"9563167e-7b47-4b29-85e8-4e299c9e10d3\" (UID: \"9563167e-7b47-4b29-85e8-4e299c9e10d3\") " Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.845634 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9563167e-7b47-4b29-85e8-4e299c9e10d3-kube-api-access-jg2ls" (OuterVolumeSpecName: "kube-api-access-jg2ls") pod "9563167e-7b47-4b29-85e8-4e299c9e10d3" (UID: "9563167e-7b47-4b29-85e8-4e299c9e10d3"). InnerVolumeSpecName "kube-api-access-jg2ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.898248 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9563167e-7b47-4b29-85e8-4e299c9e10d3" (UID: "9563167e-7b47-4b29-85e8-4e299c9e10d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.912529 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9563167e-7b47-4b29-85e8-4e299c9e10d3" (UID: "9563167e-7b47-4b29-85e8-4e299c9e10d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.923173 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-config" (OuterVolumeSpecName: "config") pod "9563167e-7b47-4b29-85e8-4e299c9e10d3" (UID: "9563167e-7b47-4b29-85e8-4e299c9e10d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.924192 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.924213 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg2ls\" (UniqueName: \"kubernetes.io/projected/9563167e-7b47-4b29-85e8-4e299c9e10d3-kube-api-access-jg2ls\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.924228 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.924237 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:24 crc kubenswrapper[4744]: I1008 09:28:24.944159 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9563167e-7b47-4b29-85e8-4e299c9e10d3" (UID: "9563167e-7b47-4b29-85e8-4e299c9e10d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:28:25 crc kubenswrapper[4744]: I1008 09:28:25.027429 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9563167e-7b47-4b29-85e8-4e299c9e10d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:25 crc kubenswrapper[4744]: I1008 09:28:25.239008 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" event={"ID":"993f71a0-3367-4a19-b0b0-b7c75b4d7e16","Type":"ContainerStarted","Data":"9da75d31e16d877276ed2b0a4eee574afad614037f0011bbf412ba3460f32bbd"} Oct 08 09:28:25 crc kubenswrapper[4744]: I1008 09:28:25.239526 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:25 crc kubenswrapper[4744]: I1008 09:28:25.248057 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" event={"ID":"9563167e-7b47-4b29-85e8-4e299c9e10d3","Type":"ContainerDied","Data":"94e8c994ab7c5a786d19d9d831f1ade5e26699e43e857001f5233a55ff14e3b0"} Oct 08 09:28:25 crc kubenswrapper[4744]: I1008 09:28:25.248125 4744 scope.go:117] "RemoveContainer" containerID="5ddfbda7ed0fecbd0c8f4e4f2496a77efbda95b751f843911a4e04da7a3c532c" Oct 08 09:28:25 crc kubenswrapper[4744]: I1008 09:28:25.248321 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7d6ff65f-28q4g" Oct 08 09:28:25 crc kubenswrapper[4744]: I1008 09:28:25.268729 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" podStartSLOduration=3.268711741 podStartE2EDuration="3.268711741s" podCreationTimestamp="2025-10-08 09:28:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:28:25.267913539 +0000 UTC m=+1000.515558848" watchObservedRunningTime="2025-10-08 09:28:25.268711741 +0000 UTC m=+1000.516356980" Oct 08 09:28:25 crc kubenswrapper[4744]: I1008 09:28:25.335815 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c7d6ff65f-28q4g"] Oct 08 09:28:25 crc kubenswrapper[4744]: I1008 09:28:25.341196 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c7d6ff65f-28q4g"] Oct 08 09:28:25 crc kubenswrapper[4744]: I1008 09:28:25.487475 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9563167e-7b47-4b29-85e8-4e299c9e10d3" path="/var/lib/kubelet/pods/9563167e-7b47-4b29-85e8-4e299c9e10d3/volumes" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.130442 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-86e2-account-create-7jnn7"] Oct 08 09:28:26 crc kubenswrapper[4744]: E1008 09:28:26.131008 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9563167e-7b47-4b29-85e8-4e299c9e10d3" containerName="init" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.131032 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9563167e-7b47-4b29-85e8-4e299c9e10d3" containerName="init" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.131241 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9563167e-7b47-4b29-85e8-4e299c9e10d3" containerName="init" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.132035 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-86e2-account-create-7jnn7" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.140784 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.147326 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-86e2-account-create-7jnn7"] Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.256765 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55k8r\" (UniqueName: \"kubernetes.io/projected/d1cf3771-9139-48f3-9d3b-9533b957aa14-kube-api-access-55k8r\") pod \"barbican-86e2-account-create-7jnn7\" (UID: \"d1cf3771-9139-48f3-9d3b-9533b957aa14\") " pod="openstack/barbican-86e2-account-create-7jnn7" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.318000 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-96af-account-create-47426"] Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.319611 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-96af-account-create-47426" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.327713 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.330291 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-96af-account-create-47426"] Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.363347 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55k8r\" (UniqueName: \"kubernetes.io/projected/d1cf3771-9139-48f3-9d3b-9533b957aa14-kube-api-access-55k8r\") pod \"barbican-86e2-account-create-7jnn7\" (UID: \"d1cf3771-9139-48f3-9d3b-9533b957aa14\") " pod="openstack/barbican-86e2-account-create-7jnn7" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.364521 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf4sn\" (UniqueName: \"kubernetes.io/projected/af798bff-4019-44a0-a0f9-f6213e2c2d64-kube-api-access-sf4sn\") pod \"cinder-96af-account-create-47426\" (UID: \"af798bff-4019-44a0-a0f9-f6213e2c2d64\") " pod="openstack/cinder-96af-account-create-47426" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.385344 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55k8r\" (UniqueName: \"kubernetes.io/projected/d1cf3771-9139-48f3-9d3b-9533b957aa14-kube-api-access-55k8r\") pod \"barbican-86e2-account-create-7jnn7\" (UID: \"d1cf3771-9139-48f3-9d3b-9533b957aa14\") " pod="openstack/barbican-86e2-account-create-7jnn7" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.458635 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-86e2-account-create-7jnn7" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.466428 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf4sn\" (UniqueName: \"kubernetes.io/projected/af798bff-4019-44a0-a0f9-f6213e2c2d64-kube-api-access-sf4sn\") pod \"cinder-96af-account-create-47426\" (UID: \"af798bff-4019-44a0-a0f9-f6213e2c2d64\") " pod="openstack/cinder-96af-account-create-47426" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.504047 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf4sn\" (UniqueName: \"kubernetes.io/projected/af798bff-4019-44a0-a0f9-f6213e2c2d64-kube-api-access-sf4sn\") pod \"cinder-96af-account-create-47426\" (UID: \"af798bff-4019-44a0-a0f9-f6213e2c2d64\") " pod="openstack/cinder-96af-account-create-47426" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.664554 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-96af-account-create-47426" Oct 08 09:28:26 crc kubenswrapper[4744]: I1008 09:28:26.837220 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-86e2-account-create-7jnn7"] Oct 08 09:28:26 crc kubenswrapper[4744]: W1008 09:28:26.839262 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1cf3771_9139_48f3_9d3b_9533b957aa14.slice/crio-dbd0fb996ca09bf565a96c83b97550ea61d4818b795fa2eb5c2f13e9d0c50f09 WatchSource:0}: Error finding container dbd0fb996ca09bf565a96c83b97550ea61d4818b795fa2eb5c2f13e9d0c50f09: Status 404 returned error can't find the container with id dbd0fb996ca09bf565a96c83b97550ea61d4818b795fa2eb5c2f13e9d0c50f09 Oct 08 09:28:27 crc kubenswrapper[4744]: I1008 09:28:27.290472 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-86e2-account-create-7jnn7" event={"ID":"d1cf3771-9139-48f3-9d3b-9533b957aa14","Type":"ContainerStarted","Data":"dbd0fb996ca09bf565a96c83b97550ea61d4818b795fa2eb5c2f13e9d0c50f09"} Oct 08 09:28:27 crc kubenswrapper[4744]: I1008 09:28:27.293222 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-96af-account-create-47426"] Oct 08 09:28:29 crc kubenswrapper[4744]: I1008 09:28:29.313437 4744 generic.go:334] "Generic (PLEG): container finished" podID="bb95350d-9809-47d8-9fd7-ae44753efe2f" containerID="798489ddd7442f1d55d1aea17083ea9445108c87809654cc94d54563aa808f0c" exitCode=0 Oct 08 09:28:29 crc kubenswrapper[4744]: I1008 09:28:29.313654 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6p2wt" event={"ID":"bb95350d-9809-47d8-9fd7-ae44753efe2f","Type":"ContainerDied","Data":"798489ddd7442f1d55d1aea17083ea9445108c87809654cc94d54563aa808f0c"} Oct 08 09:28:29 crc kubenswrapper[4744]: W1008 09:28:29.364127 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf798bff_4019_44a0_a0f9_f6213e2c2d64.slice/crio-911627240b8a4d869d9a86521d2326da16a6c5aafbbd1d9be9757a20bfa45d2b WatchSource:0}: Error finding container 911627240b8a4d869d9a86521d2326da16a6c5aafbbd1d9be9757a20bfa45d2b: Status 404 returned error can't find the container with id 911627240b8a4d869d9a86521d2326da16a6c5aafbbd1d9be9757a20bfa45d2b Oct 08 09:28:30 crc kubenswrapper[4744]: I1008 09:28:30.335123 4744 generic.go:334] "Generic (PLEG): container finished" podID="d1cf3771-9139-48f3-9d3b-9533b957aa14" containerID="9dbaad1a6c92ea0a7e75066ec1bcb443c6516f24254fa7758c499382e3c376a3" exitCode=0 Oct 08 09:28:30 crc kubenswrapper[4744]: I1008 09:28:30.335236 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-86e2-account-create-7jnn7" event={"ID":"d1cf3771-9139-48f3-9d3b-9533b957aa14","Type":"ContainerDied","Data":"9dbaad1a6c92ea0a7e75066ec1bcb443c6516f24254fa7758c499382e3c376a3"} Oct 08 09:28:30 crc kubenswrapper[4744]: I1008 09:28:30.342011 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-96af-account-create-47426" event={"ID":"af798bff-4019-44a0-a0f9-f6213e2c2d64","Type":"ContainerStarted","Data":"911627240b8a4d869d9a86521d2326da16a6c5aafbbd1d9be9757a20bfa45d2b"} Oct 08 09:28:31 crc kubenswrapper[4744]: I1008 09:28:31.937813 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:31 crc kubenswrapper[4744]: I1008 09:28:31.958742 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-86e2-account-create-7jnn7" Oct 08 09:28:31 crc kubenswrapper[4744]: I1008 09:28:31.986116 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-credential-keys\") pod \"bb95350d-9809-47d8-9fd7-ae44753efe2f\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " Oct 08 09:28:31 crc kubenswrapper[4744]: I1008 09:28:31.986253 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-config-data\") pod \"bb95350d-9809-47d8-9fd7-ae44753efe2f\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " Oct 08 09:28:31 crc kubenswrapper[4744]: I1008 09:28:31.986319 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-fernet-keys\") pod \"bb95350d-9809-47d8-9fd7-ae44753efe2f\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " Oct 08 09:28:31 crc kubenswrapper[4744]: I1008 09:28:31.986393 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mjqt\" (UniqueName: \"kubernetes.io/projected/bb95350d-9809-47d8-9fd7-ae44753efe2f-kube-api-access-6mjqt\") pod \"bb95350d-9809-47d8-9fd7-ae44753efe2f\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " Oct 08 09:28:31 crc kubenswrapper[4744]: I1008 09:28:31.986489 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-scripts\") pod \"bb95350d-9809-47d8-9fd7-ae44753efe2f\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " Oct 08 09:28:31 crc kubenswrapper[4744]: I1008 09:28:31.986542 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55k8r\" (UniqueName: \"kubernetes.io/projected/d1cf3771-9139-48f3-9d3b-9533b957aa14-kube-api-access-55k8r\") pod \"d1cf3771-9139-48f3-9d3b-9533b957aa14\" (UID: \"d1cf3771-9139-48f3-9d3b-9533b957aa14\") " Oct 08 09:28:31 crc kubenswrapper[4744]: I1008 09:28:31.986601 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-combined-ca-bundle\") pod \"bb95350d-9809-47d8-9fd7-ae44753efe2f\" (UID: \"bb95350d-9809-47d8-9fd7-ae44753efe2f\") " Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.003942 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1cf3771-9139-48f3-9d3b-9533b957aa14-kube-api-access-55k8r" (OuterVolumeSpecName: "kube-api-access-55k8r") pod "d1cf3771-9139-48f3-9d3b-9533b957aa14" (UID: "d1cf3771-9139-48f3-9d3b-9533b957aa14"). InnerVolumeSpecName "kube-api-access-55k8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.005063 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb95350d-9809-47d8-9fd7-ae44753efe2f-kube-api-access-6mjqt" (OuterVolumeSpecName: "kube-api-access-6mjqt") pod "bb95350d-9809-47d8-9fd7-ae44753efe2f" (UID: "bb95350d-9809-47d8-9fd7-ae44753efe2f"). InnerVolumeSpecName "kube-api-access-6mjqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.007150 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "bb95350d-9809-47d8-9fd7-ae44753efe2f" (UID: "bb95350d-9809-47d8-9fd7-ae44753efe2f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.017027 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bb95350d-9809-47d8-9fd7-ae44753efe2f" (UID: "bb95350d-9809-47d8-9fd7-ae44753efe2f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.023575 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-scripts" (OuterVolumeSpecName: "scripts") pod "bb95350d-9809-47d8-9fd7-ae44753efe2f" (UID: "bb95350d-9809-47d8-9fd7-ae44753efe2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.048446 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-config-data" (OuterVolumeSpecName: "config-data") pod "bb95350d-9809-47d8-9fd7-ae44753efe2f" (UID: "bb95350d-9809-47d8-9fd7-ae44753efe2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.055781 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb95350d-9809-47d8-9fd7-ae44753efe2f" (UID: "bb95350d-9809-47d8-9fd7-ae44753efe2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.088414 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.088461 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.088471 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mjqt\" (UniqueName: \"kubernetes.io/projected/bb95350d-9809-47d8-9fd7-ae44753efe2f-kube-api-access-6mjqt\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.088483 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.088494 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55k8r\" (UniqueName: \"kubernetes.io/projected/d1cf3771-9139-48f3-9d3b-9533b957aa14-kube-api-access-55k8r\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.088501 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.088510 4744 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bb95350d-9809-47d8-9fd7-ae44753efe2f-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.373603 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6p2wt" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.373614 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6p2wt" event={"ID":"bb95350d-9809-47d8-9fd7-ae44753efe2f","Type":"ContainerDied","Data":"aecd452ff4b9b42a52af9b1cf664ab56f7b6c6a088294b1d799d6af744fd3350"} Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.373691 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aecd452ff4b9b42a52af9b1cf664ab56f7b6c6a088294b1d799d6af744fd3350" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.383530 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-86e2-account-create-7jnn7" event={"ID":"d1cf3771-9139-48f3-9d3b-9533b957aa14","Type":"ContainerDied","Data":"dbd0fb996ca09bf565a96c83b97550ea61d4818b795fa2eb5c2f13e9d0c50f09"} Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.383585 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbd0fb996ca09bf565a96c83b97550ea61d4818b795fa2eb5c2f13e9d0c50f09" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.383627 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-86e2-account-create-7jnn7" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.701599 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.795610 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq"] Oct 08 09:28:32 crc kubenswrapper[4744]: I1008 09:28:32.796053 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" podUID="395ed1f9-be4b-482a-ba22-7623eecb44d6" containerName="dnsmasq-dns" containerID="cri-o://0e72b9aa009c9fbe7c47cf736ce7c073904482761dba03f427c8bb92ab81380e" gracePeriod=10 Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.075814 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6p2wt"] Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.088417 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6p2wt"] Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.173650 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rgx4v"] Oct 08 09:28:33 crc kubenswrapper[4744]: E1008 09:28:33.175696 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1cf3771-9139-48f3-9d3b-9533b957aa14" containerName="mariadb-account-create" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.175724 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1cf3771-9139-48f3-9d3b-9533b957aa14" containerName="mariadb-account-create" Oct 08 09:28:33 crc kubenswrapper[4744]: E1008 09:28:33.175744 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb95350d-9809-47d8-9fd7-ae44753efe2f" containerName="keystone-bootstrap" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.175754 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb95350d-9809-47d8-9fd7-ae44753efe2f" containerName="keystone-bootstrap" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.176057 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb95350d-9809-47d8-9fd7-ae44753efe2f" containerName="keystone-bootstrap" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.176088 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1cf3771-9139-48f3-9d3b-9533b957aa14" containerName="mariadb-account-create" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.177054 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.181209 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.181248 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.181332 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dkhdj" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.181654 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.215939 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rgx4v"] Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.249778 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-fernet-keys\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.249862 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-config-data\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.249901 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sbmk\" (UniqueName: \"kubernetes.io/projected/c6aacf4c-187c-4d2c-a740-09514e340744-kube-api-access-6sbmk\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.250013 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-credential-keys\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.250198 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-combined-ca-bundle\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.250347 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-scripts\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.352575 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-fernet-keys\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.352646 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-config-data\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.352691 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sbmk\" (UniqueName: \"kubernetes.io/projected/c6aacf4c-187c-4d2c-a740-09514e340744-kube-api-access-6sbmk\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.352734 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-credential-keys\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.352788 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-combined-ca-bundle\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.352828 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-scripts\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.361552 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-credential-keys\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.363221 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-config-data\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.363931 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-fernet-keys\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.365964 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-scripts\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.370922 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-combined-ca-bundle\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.377123 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sbmk\" (UniqueName: \"kubernetes.io/projected/c6aacf4c-187c-4d2c-a740-09514e340744-kube-api-access-6sbmk\") pod \"keystone-bootstrap-rgx4v\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.408803 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-g7tp7" event={"ID":"a6a227c0-ae15-427c-9557-72be160cb20c","Type":"ContainerStarted","Data":"ea4971b0bcfa3ea8271b5f449d7ac39af65ed01983276a03216c201f800f3d49"} Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.415465 4744 generic.go:334] "Generic (PLEG): container finished" podID="af798bff-4019-44a0-a0f9-f6213e2c2d64" containerID="559c32c64a550f3820d5cdcb0aff75cb41af4b0add66caaf9585bbc2d1a8587a" exitCode=0 Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.415629 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-96af-account-create-47426" event={"ID":"af798bff-4019-44a0-a0f9-f6213e2c2d64","Type":"ContainerDied","Data":"559c32c64a550f3820d5cdcb0aff75cb41af4b0add66caaf9585bbc2d1a8587a"} Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.425968 4744 generic.go:334] "Generic (PLEG): container finished" podID="395ed1f9-be4b-482a-ba22-7623eecb44d6" containerID="0e72b9aa009c9fbe7c47cf736ce7c073904482761dba03f427c8bb92ab81380e" exitCode=0 Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.426047 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" event={"ID":"395ed1f9-be4b-482a-ba22-7623eecb44d6","Type":"ContainerDied","Data":"0e72b9aa009c9fbe7c47cf736ce7c073904482761dba03f427c8bb92ab81380e"} Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.426218 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" event={"ID":"395ed1f9-be4b-482a-ba22-7623eecb44d6","Type":"ContainerDied","Data":"2f724d01479c2ef3e08ef186069e9b0339793c864a857ad317b3b17d3c3b73bb"} Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.426239 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f724d01479c2ef3e08ef186069e9b0339793c864a857ad317b3b17d3c3b73bb" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.428615 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04474f89-9bb4-4bb7-aef3-66d519650a4b","Type":"ContainerStarted","Data":"3aba057c1c76a344474b5d9f3b1483ae61f5c2ffce0b161ab7decf06b319e4de"} Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.442968 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-g7tp7" podStartSLOduration=2.531355049 podStartE2EDuration="11.442948914s" podCreationTimestamp="2025-10-08 09:28:22 +0000 UTC" firstStartedPulling="2025-10-08 09:28:23.494626823 +0000 UTC m=+998.742272062" lastFinishedPulling="2025-10-08 09:28:32.406220678 +0000 UTC m=+1007.653865927" observedRunningTime="2025-10-08 09:28:33.436786919 +0000 UTC m=+1008.684432158" watchObservedRunningTime="2025-10-08 09:28:33.442948914 +0000 UTC m=+1008.690594153" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.481080 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.482871 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb95350d-9809-47d8-9fd7-ae44753efe2f" path="/var/lib/kubelet/pods/bb95350d-9809-47d8-9fd7-ae44753efe2f/volumes" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.515116 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.555642 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-config\") pod \"395ed1f9-be4b-482a-ba22-7623eecb44d6\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.555698 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5mxg\" (UniqueName: \"kubernetes.io/projected/395ed1f9-be4b-482a-ba22-7623eecb44d6-kube-api-access-x5mxg\") pod \"395ed1f9-be4b-482a-ba22-7623eecb44d6\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.555910 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-ovsdbserver-nb\") pod \"395ed1f9-be4b-482a-ba22-7623eecb44d6\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.556056 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-dns-svc\") pod \"395ed1f9-be4b-482a-ba22-7623eecb44d6\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.556100 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-ovsdbserver-sb\") pod \"395ed1f9-be4b-482a-ba22-7623eecb44d6\" (UID: \"395ed1f9-be4b-482a-ba22-7623eecb44d6\") " Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.566268 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/395ed1f9-be4b-482a-ba22-7623eecb44d6-kube-api-access-x5mxg" (OuterVolumeSpecName: "kube-api-access-x5mxg") pod "395ed1f9-be4b-482a-ba22-7623eecb44d6" (UID: "395ed1f9-be4b-482a-ba22-7623eecb44d6"). InnerVolumeSpecName "kube-api-access-x5mxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.624664 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "395ed1f9-be4b-482a-ba22-7623eecb44d6" (UID: "395ed1f9-be4b-482a-ba22-7623eecb44d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.638509 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "395ed1f9-be4b-482a-ba22-7623eecb44d6" (UID: "395ed1f9-be4b-482a-ba22-7623eecb44d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.646061 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "395ed1f9-be4b-482a-ba22-7623eecb44d6" (UID: "395ed1f9-be4b-482a-ba22-7623eecb44d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.646617 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-config" (OuterVolumeSpecName: "config") pod "395ed1f9-be4b-482a-ba22-7623eecb44d6" (UID: "395ed1f9-be4b-482a-ba22-7623eecb44d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.658718 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.658754 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.658766 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.658779 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5mxg\" (UniqueName: \"kubernetes.io/projected/395ed1f9-be4b-482a-ba22-7623eecb44d6-kube-api-access-x5mxg\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:33 crc kubenswrapper[4744]: I1008 09:28:33.658788 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/395ed1f9-be4b-482a-ba22-7623eecb44d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:34 crc kubenswrapper[4744]: I1008 09:28:34.166561 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rgx4v"] Oct 08 09:28:34 crc kubenswrapper[4744]: W1008 09:28:34.175120 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6aacf4c_187c_4d2c_a740_09514e340744.slice/crio-088649b5d0d594a5868f8070960c95f8a47a94e1a0f67e05ef30c109ecd6e13b WatchSource:0}: Error finding container 088649b5d0d594a5868f8070960c95f8a47a94e1a0f67e05ef30c109ecd6e13b: Status 404 returned error can't find the container with id 088649b5d0d594a5868f8070960c95f8a47a94e1a0f67e05ef30c109ecd6e13b Oct 08 09:28:34 crc kubenswrapper[4744]: I1008 09:28:34.451304 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq" Oct 08 09:28:34 crc kubenswrapper[4744]: I1008 09:28:34.451264 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rgx4v" event={"ID":"c6aacf4c-187c-4d2c-a740-09514e340744","Type":"ContainerStarted","Data":"088649b5d0d594a5868f8070960c95f8a47a94e1a0f67e05ef30c109ecd6e13b"} Oct 08 09:28:34 crc kubenswrapper[4744]: I1008 09:28:34.499679 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq"] Oct 08 09:28:34 crc kubenswrapper[4744]: I1008 09:28:34.505843 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b5cb8c7cf-9gdbq"] Oct 08 09:28:35 crc kubenswrapper[4744]: I1008 09:28:35.329762 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-96af-account-create-47426" Oct 08 09:28:35 crc kubenswrapper[4744]: I1008 09:28:35.480228 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="395ed1f9-be4b-482a-ba22-7623eecb44d6" path="/var/lib/kubelet/pods/395ed1f9-be4b-482a-ba22-7623eecb44d6/volumes" Oct 08 09:28:35 crc kubenswrapper[4744]: I1008 09:28:35.495004 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rgx4v" event={"ID":"c6aacf4c-187c-4d2c-a740-09514e340744","Type":"ContainerStarted","Data":"60a038896f7a4eac5aceca441989984b484b422fb7dedc2d4488a50fb6b27c9f"} Oct 08 09:28:35 crc kubenswrapper[4744]: I1008 09:28:35.497335 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-96af-account-create-47426" event={"ID":"af798bff-4019-44a0-a0f9-f6213e2c2d64","Type":"ContainerDied","Data":"911627240b8a4d869d9a86521d2326da16a6c5aafbbd1d9be9757a20bfa45d2b"} Oct 08 09:28:35 crc kubenswrapper[4744]: I1008 09:28:35.497404 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="911627240b8a4d869d9a86521d2326da16a6c5aafbbd1d9be9757a20bfa45d2b" Oct 08 09:28:35 crc kubenswrapper[4744]: I1008 09:28:35.497531 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-96af-account-create-47426" Oct 08 09:28:35 crc kubenswrapper[4744]: I1008 09:28:35.498751 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf4sn\" (UniqueName: \"kubernetes.io/projected/af798bff-4019-44a0-a0f9-f6213e2c2d64-kube-api-access-sf4sn\") pod \"af798bff-4019-44a0-a0f9-f6213e2c2d64\" (UID: \"af798bff-4019-44a0-a0f9-f6213e2c2d64\") " Oct 08 09:28:35 crc kubenswrapper[4744]: I1008 09:28:35.506113 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af798bff-4019-44a0-a0f9-f6213e2c2d64-kube-api-access-sf4sn" (OuterVolumeSpecName: "kube-api-access-sf4sn") pod "af798bff-4019-44a0-a0f9-f6213e2c2d64" (UID: "af798bff-4019-44a0-a0f9-f6213e2c2d64"). InnerVolumeSpecName "kube-api-access-sf4sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:35 crc kubenswrapper[4744]: I1008 09:28:35.506180 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04474f89-9bb4-4bb7-aef3-66d519650a4b","Type":"ContainerStarted","Data":"94829c949cfcd7db8c18e78499cecb7a0bd2e2cae60407e426135cc749ea730f"} Oct 08 09:28:35 crc kubenswrapper[4744]: I1008 09:28:35.523860 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rgx4v" podStartSLOduration=2.523796538 podStartE2EDuration="2.523796538s" podCreationTimestamp="2025-10-08 09:28:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:28:35.511619851 +0000 UTC m=+1010.759265110" watchObservedRunningTime="2025-10-08 09:28:35.523796538 +0000 UTC m=+1010.771441787" Oct 08 09:28:35 crc kubenswrapper[4744]: I1008 09:28:35.600899 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf4sn\" (UniqueName: \"kubernetes.io/projected/af798bff-4019-44a0-a0f9-f6213e2c2d64-kube-api-access-sf4sn\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.386197 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-bxh5q"] Oct 08 09:28:36 crc kubenswrapper[4744]: E1008 09:28:36.387506 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af798bff-4019-44a0-a0f9-f6213e2c2d64" containerName="mariadb-account-create" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.387521 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="af798bff-4019-44a0-a0f9-f6213e2c2d64" containerName="mariadb-account-create" Oct 08 09:28:36 crc kubenswrapper[4744]: E1008 09:28:36.387552 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="395ed1f9-be4b-482a-ba22-7623eecb44d6" containerName="dnsmasq-dns" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.387558 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="395ed1f9-be4b-482a-ba22-7623eecb44d6" containerName="dnsmasq-dns" Oct 08 09:28:36 crc kubenswrapper[4744]: E1008 09:28:36.387573 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="395ed1f9-be4b-482a-ba22-7623eecb44d6" containerName="init" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.387579 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="395ed1f9-be4b-482a-ba22-7623eecb44d6" containerName="init" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.387733 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="af798bff-4019-44a0-a0f9-f6213e2c2d64" containerName="mariadb-account-create" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.387751 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="395ed1f9-be4b-482a-ba22-7623eecb44d6" containerName="dnsmasq-dns" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.388306 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.391679 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-x4jsh" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.392860 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.419055 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-bxh5q"] Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.522593 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rcb8\" (UniqueName: \"kubernetes.io/projected/4f301473-6a0d-4b3d-936d-9078c6ea6c54-kube-api-access-6rcb8\") pod \"barbican-db-sync-bxh5q\" (UID: \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\") " pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.522686 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4f301473-6a0d-4b3d-936d-9078c6ea6c54-db-sync-config-data\") pod \"barbican-db-sync-bxh5q\" (UID: \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\") " pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.523058 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f301473-6a0d-4b3d-936d-9078c6ea6c54-combined-ca-bundle\") pod \"barbican-db-sync-bxh5q\" (UID: \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\") " pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.527184 4744 generic.go:334] "Generic (PLEG): container finished" podID="a6a227c0-ae15-427c-9557-72be160cb20c" containerID="ea4971b0bcfa3ea8271b5f449d7ac39af65ed01983276a03216c201f800f3d49" exitCode=0 Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.528246 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-g7tp7" event={"ID":"a6a227c0-ae15-427c-9557-72be160cb20c","Type":"ContainerDied","Data":"ea4971b0bcfa3ea8271b5f449d7ac39af65ed01983276a03216c201f800f3d49"} Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.624491 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f301473-6a0d-4b3d-936d-9078c6ea6c54-combined-ca-bundle\") pod \"barbican-db-sync-bxh5q\" (UID: \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\") " pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.624626 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rcb8\" (UniqueName: \"kubernetes.io/projected/4f301473-6a0d-4b3d-936d-9078c6ea6c54-kube-api-access-6rcb8\") pod \"barbican-db-sync-bxh5q\" (UID: \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\") " pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.624671 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4f301473-6a0d-4b3d-936d-9078c6ea6c54-db-sync-config-data\") pod \"barbican-db-sync-bxh5q\" (UID: \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\") " pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.642748 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f301473-6a0d-4b3d-936d-9078c6ea6c54-combined-ca-bundle\") pod \"barbican-db-sync-bxh5q\" (UID: \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\") " pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.644177 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4f301473-6a0d-4b3d-936d-9078c6ea6c54-db-sync-config-data\") pod \"barbican-db-sync-bxh5q\" (UID: \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\") " pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.652114 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rcb8\" (UniqueName: \"kubernetes.io/projected/4f301473-6a0d-4b3d-936d-9078c6ea6c54-kube-api-access-6rcb8\") pod \"barbican-db-sync-bxh5q\" (UID: \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\") " pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:28:36 crc kubenswrapper[4744]: I1008 09:28:36.718449 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:28:37 crc kubenswrapper[4744]: I1008 09:28:37.290015 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-bxh5q"] Oct 08 09:28:37 crc kubenswrapper[4744]: I1008 09:28:37.539184 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bxh5q" event={"ID":"4f301473-6a0d-4b3d-936d-9078c6ea6c54","Type":"ContainerStarted","Data":"4a30e8f52bd75afe18f8eb831d1baa6225bbe058d2ad0254a5697ff99423b6d3"} Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.010324 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.162653 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-combined-ca-bundle\") pod \"a6a227c0-ae15-427c-9557-72be160cb20c\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.163261 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-scripts\") pod \"a6a227c0-ae15-427c-9557-72be160cb20c\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.163336 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6gsz\" (UniqueName: \"kubernetes.io/projected/a6a227c0-ae15-427c-9557-72be160cb20c-kube-api-access-r6gsz\") pod \"a6a227c0-ae15-427c-9557-72be160cb20c\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.163426 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6a227c0-ae15-427c-9557-72be160cb20c-logs\") pod \"a6a227c0-ae15-427c-9557-72be160cb20c\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.163456 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-config-data\") pod \"a6a227c0-ae15-427c-9557-72be160cb20c\" (UID: \"a6a227c0-ae15-427c-9557-72be160cb20c\") " Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.163962 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6a227c0-ae15-427c-9557-72be160cb20c-logs" (OuterVolumeSpecName: "logs") pod "a6a227c0-ae15-427c-9557-72be160cb20c" (UID: "a6a227c0-ae15-427c-9557-72be160cb20c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.164225 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6a227c0-ae15-427c-9557-72be160cb20c-logs\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.170166 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-scripts" (OuterVolumeSpecName: "scripts") pod "a6a227c0-ae15-427c-9557-72be160cb20c" (UID: "a6a227c0-ae15-427c-9557-72be160cb20c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.175353 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6a227c0-ae15-427c-9557-72be160cb20c-kube-api-access-r6gsz" (OuterVolumeSpecName: "kube-api-access-r6gsz") pod "a6a227c0-ae15-427c-9557-72be160cb20c" (UID: "a6a227c0-ae15-427c-9557-72be160cb20c"). InnerVolumeSpecName "kube-api-access-r6gsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.199435 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6a227c0-ae15-427c-9557-72be160cb20c" (UID: "a6a227c0-ae15-427c-9557-72be160cb20c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.209475 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-config-data" (OuterVolumeSpecName: "config-data") pod "a6a227c0-ae15-427c-9557-72be160cb20c" (UID: "a6a227c0-ae15-427c-9557-72be160cb20c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.266187 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.266244 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.266266 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6gsz\" (UniqueName: \"kubernetes.io/projected/a6a227c0-ae15-427c-9557-72be160cb20c-kube-api-access-r6gsz\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.266284 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6a227c0-ae15-427c-9557-72be160cb20c-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.554314 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-g7tp7" event={"ID":"a6a227c0-ae15-427c-9557-72be160cb20c","Type":"ContainerDied","Data":"07b1b938882f375d48c1d59963a929e7d2e1d8b53df1dd18de562f8e05e5f055"} Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.554406 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07b1b938882f375d48c1d59963a929e7d2e1d8b53df1dd18de562f8e05e5f055" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.554501 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-g7tp7" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.557559 4744 generic.go:334] "Generic (PLEG): container finished" podID="c6aacf4c-187c-4d2c-a740-09514e340744" containerID="60a038896f7a4eac5aceca441989984b484b422fb7dedc2d4488a50fb6b27c9f" exitCode=0 Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.557613 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rgx4v" event={"ID":"c6aacf4c-187c-4d2c-a740-09514e340744","Type":"ContainerDied","Data":"60a038896f7a4eac5aceca441989984b484b422fb7dedc2d4488a50fb6b27c9f"} Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.678588 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7f48465966-bh5rl"] Oct 08 09:28:38 crc kubenswrapper[4744]: E1008 09:28:38.679132 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6a227c0-ae15-427c-9557-72be160cb20c" containerName="placement-db-sync" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.679155 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6a227c0-ae15-427c-9557-72be160cb20c" containerName="placement-db-sync" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.679343 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6a227c0-ae15-427c-9557-72be160cb20c" containerName="placement-db-sync" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.692590 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.698335 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.698796 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.699162 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.699428 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.699641 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-89lxn" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.711274 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7f48465966-bh5rl"] Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.784838 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-config-data\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.784895 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxfc7\" (UniqueName: \"kubernetes.io/projected/79381d24-0ce2-46ee-ae85-6fdbcbbae230-kube-api-access-gxfc7\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.784940 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-internal-tls-certs\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.784969 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-scripts\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.785042 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79381d24-0ce2-46ee-ae85-6fdbcbbae230-logs\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.785090 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-combined-ca-bundle\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.785117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-public-tls-certs\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.887473 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-public-tls-certs\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.887588 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-config-data\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.887612 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxfc7\" (UniqueName: \"kubernetes.io/projected/79381d24-0ce2-46ee-ae85-6fdbcbbae230-kube-api-access-gxfc7\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.887656 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-internal-tls-certs\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.887685 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-scripts\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.887781 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79381d24-0ce2-46ee-ae85-6fdbcbbae230-logs\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.887832 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-combined-ca-bundle\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.888693 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79381d24-0ce2-46ee-ae85-6fdbcbbae230-logs\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.894324 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-public-tls-certs\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.915073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-scripts\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.916382 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-combined-ca-bundle\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.917200 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-internal-tls-certs\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.924164 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79381d24-0ce2-46ee-ae85-6fdbcbbae230-config-data\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:38 crc kubenswrapper[4744]: I1008 09:28:38.930160 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxfc7\" (UniqueName: \"kubernetes.io/projected/79381d24-0ce2-46ee-ae85-6fdbcbbae230-kube-api-access-gxfc7\") pod \"placement-7f48465966-bh5rl\" (UID: \"79381d24-0ce2-46ee-ae85-6fdbcbbae230\") " pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:39 crc kubenswrapper[4744]: I1008 09:28:39.016760 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:39 crc kubenswrapper[4744]: I1008 09:28:39.771223 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7f48465966-bh5rl"] Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.589557 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7f48465966-bh5rl" event={"ID":"79381d24-0ce2-46ee-ae85-6fdbcbbae230","Type":"ContainerStarted","Data":"c8761d734a7eb554f3327960ea7041089ededc1559de83def04a91ff475c53f1"} Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.769357 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-q99xn"] Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.770778 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.776081 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.776675 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.776667 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-q2q4j" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.788819 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-q99xn"] Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.879400 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stp2z\" (UniqueName: \"kubernetes.io/projected/f0779920-88cf-4475-9c77-5de26f893b6b-kube-api-access-stp2z\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.879492 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0779920-88cf-4475-9c77-5de26f893b6b-etc-machine-id\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.879538 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-combined-ca-bundle\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.879576 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-config-data\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.879659 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-scripts\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.879742 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-db-sync-config-data\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.988835 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-db-sync-config-data\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.988922 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stp2z\" (UniqueName: \"kubernetes.io/projected/f0779920-88cf-4475-9c77-5de26f893b6b-kube-api-access-stp2z\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.988954 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0779920-88cf-4475-9c77-5de26f893b6b-etc-machine-id\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.988986 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-combined-ca-bundle\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.989012 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-config-data\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.989041 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-scripts\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.989287 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0779920-88cf-4475-9c77-5de26f893b6b-etc-machine-id\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:41 crc kubenswrapper[4744]: I1008 09:28:41.994423 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-db-sync-config-data\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:42 crc kubenswrapper[4744]: I1008 09:28:42.025219 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-combined-ca-bundle\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:42 crc kubenswrapper[4744]: I1008 09:28:42.025532 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-scripts\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:42 crc kubenswrapper[4744]: I1008 09:28:42.049515 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stp2z\" (UniqueName: \"kubernetes.io/projected/f0779920-88cf-4475-9c77-5de26f893b6b-kube-api-access-stp2z\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:42 crc kubenswrapper[4744]: I1008 09:28:42.051559 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-config-data\") pod \"cinder-db-sync-q99xn\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:42 crc kubenswrapper[4744]: I1008 09:28:42.099144 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-q99xn" Oct 08 09:28:44 crc kubenswrapper[4744]: I1008 09:28:44.809579 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:44 crc kubenswrapper[4744]: I1008 09:28:44.962104 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-credential-keys\") pod \"c6aacf4c-187c-4d2c-a740-09514e340744\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " Oct 08 09:28:44 crc kubenswrapper[4744]: I1008 09:28:44.962160 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-config-data\") pod \"c6aacf4c-187c-4d2c-a740-09514e340744\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " Oct 08 09:28:44 crc kubenswrapper[4744]: I1008 09:28:44.962185 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-scripts\") pod \"c6aacf4c-187c-4d2c-a740-09514e340744\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " Oct 08 09:28:44 crc kubenswrapper[4744]: I1008 09:28:44.962472 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-combined-ca-bundle\") pod \"c6aacf4c-187c-4d2c-a740-09514e340744\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " Oct 08 09:28:44 crc kubenswrapper[4744]: I1008 09:28:44.962559 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sbmk\" (UniqueName: \"kubernetes.io/projected/c6aacf4c-187c-4d2c-a740-09514e340744-kube-api-access-6sbmk\") pod \"c6aacf4c-187c-4d2c-a740-09514e340744\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " Oct 08 09:28:44 crc kubenswrapper[4744]: I1008 09:28:44.962765 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-fernet-keys\") pod \"c6aacf4c-187c-4d2c-a740-09514e340744\" (UID: \"c6aacf4c-187c-4d2c-a740-09514e340744\") " Oct 08 09:28:44 crc kubenswrapper[4744]: I1008 09:28:44.970218 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6aacf4c-187c-4d2c-a740-09514e340744-kube-api-access-6sbmk" (OuterVolumeSpecName: "kube-api-access-6sbmk") pod "c6aacf4c-187c-4d2c-a740-09514e340744" (UID: "c6aacf4c-187c-4d2c-a740-09514e340744"). InnerVolumeSpecName "kube-api-access-6sbmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:44 crc kubenswrapper[4744]: I1008 09:28:44.970282 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c6aacf4c-187c-4d2c-a740-09514e340744" (UID: "c6aacf4c-187c-4d2c-a740-09514e340744"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:44 crc kubenswrapper[4744]: I1008 09:28:44.975696 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-scripts" (OuterVolumeSpecName: "scripts") pod "c6aacf4c-187c-4d2c-a740-09514e340744" (UID: "c6aacf4c-187c-4d2c-a740-09514e340744"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:44 crc kubenswrapper[4744]: I1008 09:28:44.976362 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c6aacf4c-187c-4d2c-a740-09514e340744" (UID: "c6aacf4c-187c-4d2c-a740-09514e340744"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:44 crc kubenswrapper[4744]: I1008 09:28:44.997344 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6aacf4c-187c-4d2c-a740-09514e340744" (UID: "c6aacf4c-187c-4d2c-a740-09514e340744"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.007811 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-config-data" (OuterVolumeSpecName: "config-data") pod "c6aacf4c-187c-4d2c-a740-09514e340744" (UID: "c6aacf4c-187c-4d2c-a740-09514e340744"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.065357 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.065440 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sbmk\" (UniqueName: \"kubernetes.io/projected/c6aacf4c-187c-4d2c-a740-09514e340744-kube-api-access-6sbmk\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.065461 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.065474 4744 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.065484 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.065495 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6aacf4c-187c-4d2c-a740-09514e340744-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.644729 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rgx4v" event={"ID":"c6aacf4c-187c-4d2c-a740-09514e340744","Type":"ContainerDied","Data":"088649b5d0d594a5868f8070960c95f8a47a94e1a0f67e05ef30c109ecd6e13b"} Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.645182 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="088649b5d0d594a5868f8070960c95f8a47a94e1a0f67e05ef30c109ecd6e13b" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.645268 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rgx4v" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.954865 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6c4d7b4588-6wcr4"] Oct 08 09:28:45 crc kubenswrapper[4744]: E1008 09:28:45.955356 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6aacf4c-187c-4d2c-a740-09514e340744" containerName="keystone-bootstrap" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.955375 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6aacf4c-187c-4d2c-a740-09514e340744" containerName="keystone-bootstrap" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.955585 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6aacf4c-187c-4d2c-a740-09514e340744" containerName="keystone-bootstrap" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.956360 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.960053 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.960358 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.960860 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.961728 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.961896 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dkhdj" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.962169 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 09:28:45 crc kubenswrapper[4744]: I1008 09:28:45.999551 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c4d7b4588-6wcr4"] Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.087987 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-fernet-keys\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.088082 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-credential-keys\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.088128 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-public-tls-certs\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.088284 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-internal-tls-certs\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.088350 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-scripts\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.088451 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-combined-ca-bundle\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.088658 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwkqh\" (UniqueName: \"kubernetes.io/projected/2af8995b-abf6-4615-bfda-f5900a1bfb2c-kube-api-access-dwkqh\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.088820 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-config-data\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.191358 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-credential-keys\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.191466 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-public-tls-certs\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.191496 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-internal-tls-certs\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.191522 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-scripts\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.191553 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-combined-ca-bundle\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.192307 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwkqh\" (UniqueName: \"kubernetes.io/projected/2af8995b-abf6-4615-bfda-f5900a1bfb2c-kube-api-access-dwkqh\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.192345 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-config-data\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.192440 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-fernet-keys\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.197976 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-public-tls-certs\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.200889 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-scripts\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.202069 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-credential-keys\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.203536 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-internal-tls-certs\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.205636 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-fernet-keys\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.207219 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-config-data\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.211512 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2af8995b-abf6-4615-bfda-f5900a1bfb2c-combined-ca-bundle\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.227817 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwkqh\" (UniqueName: \"kubernetes.io/projected/2af8995b-abf6-4615-bfda-f5900a1bfb2c-kube-api-access-dwkqh\") pod \"keystone-6c4d7b4588-6wcr4\" (UID: \"2af8995b-abf6-4615-bfda-f5900a1bfb2c\") " pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:46 crc kubenswrapper[4744]: I1008 09:28:46.278529 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:49 crc kubenswrapper[4744]: I1008 09:28:49.592815 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c4d7b4588-6wcr4"] Oct 08 09:28:49 crc kubenswrapper[4744]: I1008 09:28:49.651452 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-q99xn"] Oct 08 09:28:49 crc kubenswrapper[4744]: W1008 09:28:49.667521 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0779920_88cf_4475_9c77_5de26f893b6b.slice/crio-74af59037c0762522316ca67cae1475bf2d618fd155f9f3f801b215101498614 WatchSource:0}: Error finding container 74af59037c0762522316ca67cae1475bf2d618fd155f9f3f801b215101498614: Status 404 returned error can't find the container with id 74af59037c0762522316ca67cae1475bf2d618fd155f9f3f801b215101498614 Oct 08 09:28:49 crc kubenswrapper[4744]: I1008 09:28:49.704222 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7f48465966-bh5rl" event={"ID":"79381d24-0ce2-46ee-ae85-6fdbcbbae230","Type":"ContainerStarted","Data":"6f8782332dbb265c1c09f3ada78623af34cfe137d661f16d2efaf07118a8ab83"} Oct 08 09:28:49 crc kubenswrapper[4744]: I1008 09:28:49.704273 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7f48465966-bh5rl" event={"ID":"79381d24-0ce2-46ee-ae85-6fdbcbbae230","Type":"ContainerStarted","Data":"15a186b10ac4c188427fd414c40881195fa0fae8e3f53e23500f77740a6fb6d4"} Oct 08 09:28:49 crc kubenswrapper[4744]: I1008 09:28:49.704675 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:49 crc kubenswrapper[4744]: I1008 09:28:49.704905 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:28:49 crc kubenswrapper[4744]: I1008 09:28:49.706231 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bxh5q" event={"ID":"4f301473-6a0d-4b3d-936d-9078c6ea6c54","Type":"ContainerStarted","Data":"59ea7fea62b38cc171ee35f64831b057b77bd883211935c7db096d4df481a75d"} Oct 08 09:28:49 crc kubenswrapper[4744]: I1008 09:28:49.743142 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-q99xn" event={"ID":"f0779920-88cf-4475-9c77-5de26f893b6b","Type":"ContainerStarted","Data":"74af59037c0762522316ca67cae1475bf2d618fd155f9f3f801b215101498614"} Oct 08 09:28:49 crc kubenswrapper[4744]: I1008 09:28:49.743943 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7f48465966-bh5rl" podStartSLOduration=11.743922857 podStartE2EDuration="11.743922857s" podCreationTimestamp="2025-10-08 09:28:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:28:49.732017428 +0000 UTC m=+1024.979662667" watchObservedRunningTime="2025-10-08 09:28:49.743922857 +0000 UTC m=+1024.991568106" Oct 08 09:28:49 crc kubenswrapper[4744]: I1008 09:28:49.760950 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c4d7b4588-6wcr4" event={"ID":"2af8995b-abf6-4615-bfda-f5900a1bfb2c","Type":"ContainerStarted","Data":"2b4b5451b97702e5ad9f714d6bc40fa0327d6ebe59538b1baa46ce803704f5a1"} Oct 08 09:28:49 crc kubenswrapper[4744]: I1008 09:28:49.770098 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-bxh5q" podStartSLOduration=2.000504217 podStartE2EDuration="13.770080062s" podCreationTimestamp="2025-10-08 09:28:36 +0000 UTC" firstStartedPulling="2025-10-08 09:28:37.31380237 +0000 UTC m=+1012.561447619" lastFinishedPulling="2025-10-08 09:28:49.083378225 +0000 UTC m=+1024.331023464" observedRunningTime="2025-10-08 09:28:49.769094905 +0000 UTC m=+1025.016740144" watchObservedRunningTime="2025-10-08 09:28:49.770080062 +0000 UTC m=+1025.017725301" Oct 08 09:28:49 crc kubenswrapper[4744]: I1008 09:28:49.772584 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04474f89-9bb4-4bb7-aef3-66d519650a4b","Type":"ContainerStarted","Data":"c4ccd4d1e8dbb1fc765510aedb9d23ab8535bb63e928a42962e2dd1da163e4db"} Oct 08 09:28:50 crc kubenswrapper[4744]: I1008 09:28:50.806558 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c4d7b4588-6wcr4" event={"ID":"2af8995b-abf6-4615-bfda-f5900a1bfb2c","Type":"ContainerStarted","Data":"fffd47fea36297dd066d4681041566ed82f9622012b504505c37acd4f598084f"} Oct 08 09:28:50 crc kubenswrapper[4744]: I1008 09:28:50.834395 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6c4d7b4588-6wcr4" podStartSLOduration=5.834360395 podStartE2EDuration="5.834360395s" podCreationTimestamp="2025-10-08 09:28:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:28:50.834143239 +0000 UTC m=+1026.081788478" watchObservedRunningTime="2025-10-08 09:28:50.834360395 +0000 UTC m=+1026.082005634" Oct 08 09:28:51 crc kubenswrapper[4744]: I1008 09:28:51.819102 4744 generic.go:334] "Generic (PLEG): container finished" podID="54a05787-2fe8-4d6d-b785-058040260352" containerID="2544b2efd54a6b3b29f077a1fd78405bc7739e74cf3844435e05fc27b230527e" exitCode=0 Oct 08 09:28:51 crc kubenswrapper[4744]: I1008 09:28:51.820770 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fxgnx" event={"ID":"54a05787-2fe8-4d6d-b785-058040260352","Type":"ContainerDied","Data":"2544b2efd54a6b3b29f077a1fd78405bc7739e74cf3844435e05fc27b230527e"} Oct 08 09:28:51 crc kubenswrapper[4744]: I1008 09:28:51.820816 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:28:52 crc kubenswrapper[4744]: I1008 09:28:52.831286 4744 generic.go:334] "Generic (PLEG): container finished" podID="4f301473-6a0d-4b3d-936d-9078c6ea6c54" containerID="59ea7fea62b38cc171ee35f64831b057b77bd883211935c7db096d4df481a75d" exitCode=0 Oct 08 09:28:52 crc kubenswrapper[4744]: I1008 09:28:52.831505 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bxh5q" event={"ID":"4f301473-6a0d-4b3d-936d-9078c6ea6c54","Type":"ContainerDied","Data":"59ea7fea62b38cc171ee35f64831b057b77bd883211935c7db096d4df481a75d"} Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.221428 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.274495 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hth5h\" (UniqueName: \"kubernetes.io/projected/54a05787-2fe8-4d6d-b785-058040260352-kube-api-access-hth5h\") pod \"54a05787-2fe8-4d6d-b785-058040260352\" (UID: \"54a05787-2fe8-4d6d-b785-058040260352\") " Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.274592 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/54a05787-2fe8-4d6d-b785-058040260352-config\") pod \"54a05787-2fe8-4d6d-b785-058040260352\" (UID: \"54a05787-2fe8-4d6d-b785-058040260352\") " Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.274626 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a05787-2fe8-4d6d-b785-058040260352-combined-ca-bundle\") pod \"54a05787-2fe8-4d6d-b785-058040260352\" (UID: \"54a05787-2fe8-4d6d-b785-058040260352\") " Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.284344 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54a05787-2fe8-4d6d-b785-058040260352-kube-api-access-hth5h" (OuterVolumeSpecName: "kube-api-access-hth5h") pod "54a05787-2fe8-4d6d-b785-058040260352" (UID: "54a05787-2fe8-4d6d-b785-058040260352"). InnerVolumeSpecName "kube-api-access-hth5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.307318 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a05787-2fe8-4d6d-b785-058040260352-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54a05787-2fe8-4d6d-b785-058040260352" (UID: "54a05787-2fe8-4d6d-b785-058040260352"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.309314 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54a05787-2fe8-4d6d-b785-058040260352-config" (OuterVolumeSpecName: "config") pod "54a05787-2fe8-4d6d-b785-058040260352" (UID: "54a05787-2fe8-4d6d-b785-058040260352"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.376201 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/54a05787-2fe8-4d6d-b785-058040260352-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.376246 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54a05787-2fe8-4d6d-b785-058040260352-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.376287 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hth5h\" (UniqueName: \"kubernetes.io/projected/54a05787-2fe8-4d6d-b785-058040260352-kube-api-access-hth5h\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.845433 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-fxgnx" event={"ID":"54a05787-2fe8-4d6d-b785-058040260352","Type":"ContainerDied","Data":"35bd34f0c504bf49b11d3634c24956712872c125a0c7b06b63afd7201a305031"} Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.845499 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35bd34f0c504bf49b11d3634c24956712872c125a0c7b06b63afd7201a305031" Oct 08 09:28:53 crc kubenswrapper[4744]: I1008 09:28:53.845449 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-fxgnx" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.201168 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d7d647849-6cdt4"] Oct 08 09:28:54 crc kubenswrapper[4744]: E1008 09:28:54.206042 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a05787-2fe8-4d6d-b785-058040260352" containerName="neutron-db-sync" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.206090 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a05787-2fe8-4d6d-b785-058040260352" containerName="neutron-db-sync" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.206526 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="54a05787-2fe8-4d6d-b785-058040260352" containerName="neutron-db-sync" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.207808 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.223723 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d7d647849-6cdt4"] Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.365171 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7f466fb8c4-5nw45"] Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.367168 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.372754 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-bp6tb" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.372775 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.373020 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.373417 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.395522 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f466fb8c4-5nw45"] Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.404775 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-ovsdbserver-sb\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.404826 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-config\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.404854 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-ovsdbserver-nb\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.404879 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-dns-svc\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.404966 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px9mb\" (UniqueName: \"kubernetes.io/projected/361e3004-e6f9-4f22-a391-ae768a43fa71-kube-api-access-px9mb\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.507262 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px9mb\" (UniqueName: \"kubernetes.io/projected/361e3004-e6f9-4f22-a391-ae768a43fa71-kube-api-access-px9mb\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.507337 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-ovndb-tls-certs\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.507383 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpjjq\" (UniqueName: \"kubernetes.io/projected/94715e15-b083-4428-95d6-d7e66990af08-kube-api-access-gpjjq\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.507426 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-combined-ca-bundle\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.507454 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-httpd-config\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.507627 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-ovsdbserver-sb\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.507778 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-config\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.507820 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-ovsdbserver-nb\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.507911 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-dns-svc\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.508151 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-config\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.509039 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-config\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.509052 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-ovsdbserver-nb\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.509314 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-dns-svc\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.509529 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-ovsdbserver-sb\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.551872 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px9mb\" (UniqueName: \"kubernetes.io/projected/361e3004-e6f9-4f22-a391-ae768a43fa71-kube-api-access-px9mb\") pod \"dnsmasq-dns-6d7d647849-6cdt4\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.610037 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-ovndb-tls-certs\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.610132 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpjjq\" (UniqueName: \"kubernetes.io/projected/94715e15-b083-4428-95d6-d7e66990af08-kube-api-access-gpjjq\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.610165 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-combined-ca-bundle\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.610193 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-httpd-config\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.610360 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-config\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.617497 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-httpd-config\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.620819 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-ovndb-tls-certs\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.624275 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-config\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.630908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-combined-ca-bundle\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.643355 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpjjq\" (UniqueName: \"kubernetes.io/projected/94715e15-b083-4428-95d6-d7e66990af08-kube-api-access-gpjjq\") pod \"neutron-7f466fb8c4-5nw45\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.707996 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:28:54 crc kubenswrapper[4744]: I1008 09:28:54.848765 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.354109 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6b4cb6945c-kjx6m"] Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.356442 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.368304 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.368571 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.375574 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6b4cb6945c-kjx6m"] Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.401016 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-internal-tls-certs\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.401174 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-combined-ca-bundle\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.401450 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-config\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.401507 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-httpd-config\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.401703 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pmsk\" (UniqueName: \"kubernetes.io/projected/d7cda27f-1e6a-4637-b647-54f6a8235ce0-kube-api-access-5pmsk\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.401856 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-ovndb-tls-certs\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.402048 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-public-tls-certs\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.504421 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-public-tls-certs\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.504600 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-internal-tls-certs\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.504626 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-combined-ca-bundle\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.504672 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-config\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.504698 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-httpd-config\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.504740 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pmsk\" (UniqueName: \"kubernetes.io/projected/d7cda27f-1e6a-4637-b647-54f6a8235ce0-kube-api-access-5pmsk\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.504798 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-ovndb-tls-certs\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.515060 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-config\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.516718 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-combined-ca-bundle\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.516753 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-internal-tls-certs\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.518051 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-httpd-config\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.525744 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-ovndb-tls-certs\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.526500 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pmsk\" (UniqueName: \"kubernetes.io/projected/d7cda27f-1e6a-4637-b647-54f6a8235ce0-kube-api-access-5pmsk\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.537664 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cda27f-1e6a-4637-b647-54f6a8235ce0-public-tls-certs\") pod \"neutron-6b4cb6945c-kjx6m\" (UID: \"d7cda27f-1e6a-4637-b647-54f6a8235ce0\") " pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:57 crc kubenswrapper[4744]: I1008 09:28:57.706553 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.246678 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.347620 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f301473-6a0d-4b3d-936d-9078c6ea6c54-combined-ca-bundle\") pod \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\" (UID: \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\") " Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.347884 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4f301473-6a0d-4b3d-936d-9078c6ea6c54-db-sync-config-data\") pod \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\" (UID: \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\") " Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.347989 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rcb8\" (UniqueName: \"kubernetes.io/projected/4f301473-6a0d-4b3d-936d-9078c6ea6c54-kube-api-access-6rcb8\") pod \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\" (UID: \"4f301473-6a0d-4b3d-936d-9078c6ea6c54\") " Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.353631 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f301473-6a0d-4b3d-936d-9078c6ea6c54-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4f301473-6a0d-4b3d-936d-9078c6ea6c54" (UID: "4f301473-6a0d-4b3d-936d-9078c6ea6c54"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.366383 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f301473-6a0d-4b3d-936d-9078c6ea6c54-kube-api-access-6rcb8" (OuterVolumeSpecName: "kube-api-access-6rcb8") pod "4f301473-6a0d-4b3d-936d-9078c6ea6c54" (UID: "4f301473-6a0d-4b3d-936d-9078c6ea6c54"). InnerVolumeSpecName "kube-api-access-6rcb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.388061 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f301473-6a0d-4b3d-936d-9078c6ea6c54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f301473-6a0d-4b3d-936d-9078c6ea6c54" (UID: "4f301473-6a0d-4b3d-936d-9078c6ea6c54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.451844 4744 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4f301473-6a0d-4b3d-936d-9078c6ea6c54-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.451884 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rcb8\" (UniqueName: \"kubernetes.io/projected/4f301473-6a0d-4b3d-936d-9078c6ea6c54-kube-api-access-6rcb8\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.451899 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f301473-6a0d-4b3d-936d-9078c6ea6c54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.948581 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bxh5q" event={"ID":"4f301473-6a0d-4b3d-936d-9078c6ea6c54","Type":"ContainerDied","Data":"4a30e8f52bd75afe18f8eb831d1baa6225bbe058d2ad0254a5697ff99423b6d3"} Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.949172 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a30e8f52bd75afe18f8eb831d1baa6225bbe058d2ad0254a5697ff99423b6d3" Oct 08 09:28:59 crc kubenswrapper[4744]: I1008 09:28:59.949293 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bxh5q" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.585069 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-8578895ff6-xg92r"] Oct 08 09:29:00 crc kubenswrapper[4744]: E1008 09:29:00.586003 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f301473-6a0d-4b3d-936d-9078c6ea6c54" containerName="barbican-db-sync" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.586027 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f301473-6a0d-4b3d-936d-9078c6ea6c54" containerName="barbican-db-sync" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.586278 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f301473-6a0d-4b3d-936d-9078c6ea6c54" containerName="barbican-db-sync" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.587476 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.591138 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.591587 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.591597 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-x4jsh" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.619629 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5ccc6f877f-j2zc9"] Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.621538 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.623763 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.627195 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-8578895ff6-xg92r"] Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.653815 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5ccc6f877f-j2zc9"] Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.671529 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ca73575-280f-43af-8c34-123a1cf83e92-config-data-custom\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.671602 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/203b8276-cafa-4b89-a68d-0f9437ee7161-logs\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.671628 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/203b8276-cafa-4b89-a68d-0f9437ee7161-config-data-custom\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.671680 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ca73575-280f-43af-8c34-123a1cf83e92-logs\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.671702 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpjnb\" (UniqueName: \"kubernetes.io/projected/4ca73575-280f-43af-8c34-123a1cf83e92-kube-api-access-xpjnb\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.671720 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca73575-280f-43af-8c34-123a1cf83e92-combined-ca-bundle\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.671741 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/203b8276-cafa-4b89-a68d-0f9437ee7161-combined-ca-bundle\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.671775 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpmhw\" (UniqueName: \"kubernetes.io/projected/203b8276-cafa-4b89-a68d-0f9437ee7161-kube-api-access-mpmhw\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.671811 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ca73575-280f-43af-8c34-123a1cf83e92-config-data\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.671827 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/203b8276-cafa-4b89-a68d-0f9437ee7161-config-data\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.719218 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d7d647849-6cdt4"] Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.737840 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5bdc4b9-8znmp"] Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.739837 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.763636 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5bdc4b9-8znmp"] Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772529 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-dns-svc\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772576 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpmhw\" (UniqueName: \"kubernetes.io/projected/203b8276-cafa-4b89-a68d-0f9437ee7161-kube-api-access-mpmhw\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772615 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/203b8276-cafa-4b89-a68d-0f9437ee7161-config-data\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772632 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ca73575-280f-43af-8c34-123a1cf83e92-config-data\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772659 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ca73575-280f-43af-8c34-123a1cf83e92-config-data-custom\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772675 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fztsc\" (UniqueName: \"kubernetes.io/projected/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-kube-api-access-fztsc\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772696 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772721 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/203b8276-cafa-4b89-a68d-0f9437ee7161-logs\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772740 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/203b8276-cafa-4b89-a68d-0f9437ee7161-config-data-custom\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772760 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-config\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772804 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ca73575-280f-43af-8c34-123a1cf83e92-logs\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772821 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpjnb\" (UniqueName: \"kubernetes.io/projected/4ca73575-280f-43af-8c34-123a1cf83e92-kube-api-access-xpjnb\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772839 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772858 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca73575-280f-43af-8c34-123a1cf83e92-combined-ca-bundle\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.772879 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/203b8276-cafa-4b89-a68d-0f9437ee7161-combined-ca-bundle\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.782120 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/203b8276-cafa-4b89-a68d-0f9437ee7161-combined-ca-bundle\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.782591 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/203b8276-cafa-4b89-a68d-0f9437ee7161-logs\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.792811 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ca73575-280f-43af-8c34-123a1cf83e92-logs\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.810095 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/203b8276-cafa-4b89-a68d-0f9437ee7161-config-data\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.811020 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/203b8276-cafa-4b89-a68d-0f9437ee7161-config-data-custom\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.816327 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ca73575-280f-43af-8c34-123a1cf83e92-config-data-custom\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.816827 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca73575-280f-43af-8c34-123a1cf83e92-combined-ca-bundle\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.816874 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ca73575-280f-43af-8c34-123a1cf83e92-config-data\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.833967 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpjnb\" (UniqueName: \"kubernetes.io/projected/4ca73575-280f-43af-8c34-123a1cf83e92-kube-api-access-xpjnb\") pod \"barbican-keystone-listener-8578895ff6-xg92r\" (UID: \"4ca73575-280f-43af-8c34-123a1cf83e92\") " pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.839875 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpmhw\" (UniqueName: \"kubernetes.io/projected/203b8276-cafa-4b89-a68d-0f9437ee7161-kube-api-access-mpmhw\") pod \"barbican-worker-5ccc6f877f-j2zc9\" (UID: \"203b8276-cafa-4b89-a68d-0f9437ee7161\") " pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.870524 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5f6598d64-fmxvg"] Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.872829 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.875284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fztsc\" (UniqueName: \"kubernetes.io/projected/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-kube-api-access-fztsc\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.875340 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.875390 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-config\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.875442 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.875472 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-dns-svc\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.876337 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-dns-svc\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.877180 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.884003 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f6598d64-fmxvg"] Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.891251 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.897053 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.897758 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-config\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.917213 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.929868 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fztsc\" (UniqueName: \"kubernetes.io/projected/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-kube-api-access-fztsc\") pod \"dnsmasq-dns-7ff5bdc4b9-8znmp\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.966072 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5ccc6f877f-j2zc9" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.976574 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-combined-ca-bundle\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.976780 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-config-data-custom\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.976953 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a50c5d4e-abd2-43bd-9c64-8248b89fb213-logs\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.977039 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-config-data\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:00 crc kubenswrapper[4744]: I1008 09:29:00.977137 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vslb9\" (UniqueName: \"kubernetes.io/projected/a50c5d4e-abd2-43bd-9c64-8248b89fb213-kube-api-access-vslb9\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:01 crc kubenswrapper[4744]: I1008 09:29:01.073205 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:01 crc kubenswrapper[4744]: I1008 09:29:01.078709 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a50c5d4e-abd2-43bd-9c64-8248b89fb213-logs\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:01 crc kubenswrapper[4744]: I1008 09:29:01.078764 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-config-data\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:01 crc kubenswrapper[4744]: I1008 09:29:01.078816 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vslb9\" (UniqueName: \"kubernetes.io/projected/a50c5d4e-abd2-43bd-9c64-8248b89fb213-kube-api-access-vslb9\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:01 crc kubenswrapper[4744]: I1008 09:29:01.078893 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-combined-ca-bundle\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:01 crc kubenswrapper[4744]: I1008 09:29:01.078923 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-config-data-custom\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:01 crc kubenswrapper[4744]: I1008 09:29:01.079161 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a50c5d4e-abd2-43bd-9c64-8248b89fb213-logs\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:01 crc kubenswrapper[4744]: I1008 09:29:01.085619 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-combined-ca-bundle\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:01 crc kubenswrapper[4744]: I1008 09:29:01.087119 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-config-data-custom\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:01 crc kubenswrapper[4744]: I1008 09:29:01.092579 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-config-data\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:01 crc kubenswrapper[4744]: I1008 09:29:01.112856 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vslb9\" (UniqueName: \"kubernetes.io/projected/a50c5d4e-abd2-43bd-9c64-8248b89fb213-kube-api-access-vslb9\") pod \"barbican-api-5f6598d64-fmxvg\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:01 crc kubenswrapper[4744]: I1008 09:29:01.271103 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.328331 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6d98c4654d-2xrfx"] Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.330528 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.334163 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.342912 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.368476 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d98c4654d-2xrfx"] Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.439823 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-public-tls-certs\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.439888 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-combined-ca-bundle\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.439977 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-config-data-custom\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.440020 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-logs\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.440062 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-config-data\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.440093 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-internal-tls-certs\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.440264 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm6nw\" (UniqueName: \"kubernetes.io/projected/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-kube-api-access-dm6nw\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.542156 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-public-tls-certs\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.542224 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-combined-ca-bundle\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.542248 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-config-data-custom\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.542280 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-logs\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.542308 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-config-data\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.542335 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-internal-tls-certs\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.542404 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm6nw\" (UniqueName: \"kubernetes.io/projected/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-kube-api-access-dm6nw\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.543904 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-logs\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.550544 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-config-data\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.558640 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-public-tls-certs\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.559323 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-combined-ca-bundle\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.559871 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-config-data-custom\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.574894 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-internal-tls-certs\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.577070 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm6nw\" (UniqueName: \"kubernetes.io/projected/ea9a2551-0abd-4bce-9785-4c8c7e11b02d-kube-api-access-dm6nw\") pod \"barbican-api-6d98c4654d-2xrfx\" (UID: \"ea9a2551-0abd-4bce-9785-4c8c7e11b02d\") " pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:03 crc kubenswrapper[4744]: I1008 09:29:03.652076 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:10 crc kubenswrapper[4744]: E1008 09:29:10.706519 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f" Oct 08 09:29:10 crc kubenswrapper[4744]: E1008 09:29:10.707574 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-stp2z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-q99xn_openstack(f0779920-88cf-4475-9c77-5de26f893b6b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 09:29:10 crc kubenswrapper[4744]: E1008 09:29:10.708797 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-q99xn" podUID="f0779920-88cf-4475-9c77-5de26f893b6b" Oct 08 09:29:10 crc kubenswrapper[4744]: I1008 09:29:10.835979 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:29:10 crc kubenswrapper[4744]: I1008 09:29:10.840453 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7f48465966-bh5rl" Oct 08 09:29:11 crc kubenswrapper[4744]: E1008 09:29:11.081134 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f\\\"\"" pod="openstack/cinder-db-sync-q99xn" podUID="f0779920-88cf-4475-9c77-5de26f893b6b" Oct 08 09:29:11 crc kubenswrapper[4744]: E1008 09:29:11.462096 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48" Oct 08 09:29:11 crc kubenswrapper[4744]: E1008 09:29:11.462295 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vrjgk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(04474f89-9bb4-4bb7-aef3-66d519650a4b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 09:29:11 crc kubenswrapper[4744]: E1008 09:29:11.463441 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" Oct 08 09:29:12 crc kubenswrapper[4744]: I1008 09:29:12.089440 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerName="ceilometer-central-agent" containerID="cri-o://3aba057c1c76a344474b5d9f3b1483ae61f5c2ffce0b161ab7decf06b319e4de" gracePeriod=30 Oct 08 09:29:12 crc kubenswrapper[4744]: I1008 09:29:12.089905 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerName="sg-core" containerID="cri-o://c4ccd4d1e8dbb1fc765510aedb9d23ab8535bb63e928a42962e2dd1da163e4db" gracePeriod=30 Oct 08 09:29:12 crc kubenswrapper[4744]: I1008 09:29:12.089962 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerName="ceilometer-notification-agent" containerID="cri-o://94829c949cfcd7db8c18e78499cecb7a0bd2e2cae60407e426135cc749ea730f" gracePeriod=30 Oct 08 09:29:12 crc kubenswrapper[4744]: I1008 09:29:12.134730 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f466fb8c4-5nw45"] Oct 08 09:29:12 crc kubenswrapper[4744]: I1008 09:29:12.326400 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6b4cb6945c-kjx6m"] Oct 08 09:29:12 crc kubenswrapper[4744]: W1008 09:29:12.336996 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7cda27f_1e6a_4637_b647_54f6a8235ce0.slice/crio-c4642dfb9bf6c065cd952f2fcfa11f8cd8d25757f9d4199f260255d3a115359c WatchSource:0}: Error finding container c4642dfb9bf6c065cd952f2fcfa11f8cd8d25757f9d4199f260255d3a115359c: Status 404 returned error can't find the container with id c4642dfb9bf6c065cd952f2fcfa11f8cd8d25757f9d4199f260255d3a115359c Oct 08 09:29:12 crc kubenswrapper[4744]: I1008 09:29:12.367596 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5ccc6f877f-j2zc9"] Oct 08 09:29:12 crc kubenswrapper[4744]: I1008 09:29:12.418218 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-8578895ff6-xg92r"] Oct 08 09:29:12 crc kubenswrapper[4744]: W1008 09:29:12.458732 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ca73575_280f_43af_8c34_123a1cf83e92.slice/crio-0ad4f26ac1ce03efd44f293835d30a01f44060f6694e76de3665369a99f59e65 WatchSource:0}: Error finding container 0ad4f26ac1ce03efd44f293835d30a01f44060f6694e76de3665369a99f59e65: Status 404 returned error can't find the container with id 0ad4f26ac1ce03efd44f293835d30a01f44060f6694e76de3665369a99f59e65 Oct 08 09:29:12 crc kubenswrapper[4744]: I1008 09:29:12.491146 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f6598d64-fmxvg"] Oct 08 09:29:12 crc kubenswrapper[4744]: I1008 09:29:12.510980 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5bdc4b9-8znmp"] Oct 08 09:29:12 crc kubenswrapper[4744]: I1008 09:29:12.540417 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d98c4654d-2xrfx"] Oct 08 09:29:12 crc kubenswrapper[4744]: I1008 09:29:12.645210 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d7d647849-6cdt4"] Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.104536 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b4cb6945c-kjx6m" event={"ID":"d7cda27f-1e6a-4637-b647-54f6a8235ce0","Type":"ContainerStarted","Data":"0d3ece8aedf136b0a62fdb7cde3064b1d92a71995a0ea595a649d7f991a25e61"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.105044 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b4cb6945c-kjx6m" event={"ID":"d7cda27f-1e6a-4637-b647-54f6a8235ce0","Type":"ContainerStarted","Data":"c4642dfb9bf6c065cd952f2fcfa11f8cd8d25757f9d4199f260255d3a115359c"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.110651 4744 generic.go:334] "Generic (PLEG): container finished" podID="8306420e-9e3d-48a6-87b9-af6c61f0d0ba" containerID="d0f22e1597065354a52bdc3836dabe667b45b928fc1bc324cf462d7cd4f0f325" exitCode=0 Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.110724 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" event={"ID":"8306420e-9e3d-48a6-87b9-af6c61f0d0ba","Type":"ContainerDied","Data":"d0f22e1597065354a52bdc3836dabe667b45b928fc1bc324cf462d7cd4f0f325"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.110749 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" event={"ID":"8306420e-9e3d-48a6-87b9-af6c61f0d0ba","Type":"ContainerStarted","Data":"b3531898dd0b6c3170e6437b94f43f0f4d0d1280bf6f9ced9fa18d2125c8f25f"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.121930 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" event={"ID":"4ca73575-280f-43af-8c34-123a1cf83e92","Type":"ContainerStarted","Data":"0ad4f26ac1ce03efd44f293835d30a01f44060f6694e76de3665369a99f59e65"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.143823 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" event={"ID":"361e3004-e6f9-4f22-a391-ae768a43fa71","Type":"ContainerStarted","Data":"34ef5894affe992d10e99cc365bb964e85e1ce767cf86344c936942fc614a7e2"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.144066 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" event={"ID":"361e3004-e6f9-4f22-a391-ae768a43fa71","Type":"ContainerStarted","Data":"60d7c42d4fef76c22f1f2f6c5091e7fa3d065ed4750145df16a1894a60f5a2cd"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.150963 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f6598d64-fmxvg" event={"ID":"a50c5d4e-abd2-43bd-9c64-8248b89fb213","Type":"ContainerStarted","Data":"0a23d118075d97f8edfe407163951087016299ba68ba0e808fe230d74f207b23"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.151135 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f6598d64-fmxvg" event={"ID":"a50c5d4e-abd2-43bd-9c64-8248b89fb213","Type":"ContainerStarted","Data":"d03420725a90694970cefe57d1af847a35d5d15a69b61b15f5410e201b065212"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.153593 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5ccc6f877f-j2zc9" event={"ID":"203b8276-cafa-4b89-a68d-0f9437ee7161","Type":"ContainerStarted","Data":"96b50f7b53c565be08bad01e965aaa3c76eeddcdb60aacbb79104d9e5571fe14"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.154654 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d98c4654d-2xrfx" event={"ID":"ea9a2551-0abd-4bce-9785-4c8c7e11b02d","Type":"ContainerStarted","Data":"2b4372ed2a8be9bc0d389d13c07888559811767da142569f340d1e1976d1ffba"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.154674 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d98c4654d-2xrfx" event={"ID":"ea9a2551-0abd-4bce-9785-4c8c7e11b02d","Type":"ContainerStarted","Data":"46bba11c2d33495b14a4f97478a5cf778a19d5ea9d39905a3a6b93eb5b83422a"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.155995 4744 generic.go:334] "Generic (PLEG): container finished" podID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerID="c4ccd4d1e8dbb1fc765510aedb9d23ab8535bb63e928a42962e2dd1da163e4db" exitCode=2 Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.156013 4744 generic.go:334] "Generic (PLEG): container finished" podID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerID="3aba057c1c76a344474b5d9f3b1483ae61f5c2ffce0b161ab7decf06b319e4de" exitCode=0 Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.156039 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04474f89-9bb4-4bb7-aef3-66d519650a4b","Type":"ContainerDied","Data":"c4ccd4d1e8dbb1fc765510aedb9d23ab8535bb63e928a42962e2dd1da163e4db"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.156054 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04474f89-9bb4-4bb7-aef3-66d519650a4b","Type":"ContainerDied","Data":"3aba057c1c76a344474b5d9f3b1483ae61f5c2ffce0b161ab7decf06b319e4de"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.179008 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f466fb8c4-5nw45" event={"ID":"94715e15-b083-4428-95d6-d7e66990af08","Type":"ContainerStarted","Data":"cd4ce2223ffbac03d43a8d5f76019aaa87e93027c60ead4fd2f9caaa285e42cb"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.179058 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f466fb8c4-5nw45" event={"ID":"94715e15-b083-4428-95d6-d7e66990af08","Type":"ContainerStarted","Data":"dbafe4df8d4eb5e4343d7baac824d5f64992310c6264c9583d609941e07ebfd7"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.179067 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f466fb8c4-5nw45" event={"ID":"94715e15-b083-4428-95d6-d7e66990af08","Type":"ContainerStarted","Data":"0b6d1d92f1dbf328db183f181ed6c7c9c3dc0b4c146b9ac4dcd3e59e65441d68"} Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.184465 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.228919 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7f466fb8c4-5nw45" podStartSLOduration=19.228897561 podStartE2EDuration="19.228897561s" podCreationTimestamp="2025-10-08 09:28:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:29:13.2266035 +0000 UTC m=+1048.474248759" watchObservedRunningTime="2025-10-08 09:29:13.228897561 +0000 UTC m=+1048.476542800" Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.763693 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.822358 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-dns-svc\") pod \"361e3004-e6f9-4f22-a391-ae768a43fa71\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.823319 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px9mb\" (UniqueName: \"kubernetes.io/projected/361e3004-e6f9-4f22-a391-ae768a43fa71-kube-api-access-px9mb\") pod \"361e3004-e6f9-4f22-a391-ae768a43fa71\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.823454 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-ovsdbserver-sb\") pod \"361e3004-e6f9-4f22-a391-ae768a43fa71\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.823733 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-config\") pod \"361e3004-e6f9-4f22-a391-ae768a43fa71\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.824117 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-ovsdbserver-nb\") pod \"361e3004-e6f9-4f22-a391-ae768a43fa71\" (UID: \"361e3004-e6f9-4f22-a391-ae768a43fa71\") " Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.831164 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/361e3004-e6f9-4f22-a391-ae768a43fa71-kube-api-access-px9mb" (OuterVolumeSpecName: "kube-api-access-px9mb") pod "361e3004-e6f9-4f22-a391-ae768a43fa71" (UID: "361e3004-e6f9-4f22-a391-ae768a43fa71"). InnerVolumeSpecName "kube-api-access-px9mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.888941 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "361e3004-e6f9-4f22-a391-ae768a43fa71" (UID: "361e3004-e6f9-4f22-a391-ae768a43fa71"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.889133 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "361e3004-e6f9-4f22-a391-ae768a43fa71" (UID: "361e3004-e6f9-4f22-a391-ae768a43fa71"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.891799 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "361e3004-e6f9-4f22-a391-ae768a43fa71" (UID: "361e3004-e6f9-4f22-a391-ae768a43fa71"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.916365 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-config" (OuterVolumeSpecName: "config") pod "361e3004-e6f9-4f22-a391-ae768a43fa71" (UID: "361e3004-e6f9-4f22-a391-ae768a43fa71"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.927356 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.927434 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px9mb\" (UniqueName: \"kubernetes.io/projected/361e3004-e6f9-4f22-a391-ae768a43fa71-kube-api-access-px9mb\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.927452 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.927466 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:13 crc kubenswrapper[4744]: I1008 09:29:13.927483 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/361e3004-e6f9-4f22-a391-ae768a43fa71-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.191532 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f6598d64-fmxvg" event={"ID":"a50c5d4e-abd2-43bd-9c64-8248b89fb213","Type":"ContainerStarted","Data":"7cd014939d983f7e9915d23c606e20670fd8733ac31eeb6cdffc4d2381684a63"} Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.193177 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.193212 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.196945 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d98c4654d-2xrfx" event={"ID":"ea9a2551-0abd-4bce-9785-4c8c7e11b02d","Type":"ContainerStarted","Data":"d49d21cf719d752bb30707fd1348fecc061643ab311d2c56d19660823dc1c62b"} Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.197110 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.199294 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b4cb6945c-kjx6m" event={"ID":"d7cda27f-1e6a-4637-b647-54f6a8235ce0","Type":"ContainerStarted","Data":"a9c03e30c0b3d952de86db4efd3e25590b99986cb1ef3c11590449a0e0c1dc2f"} Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.199764 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.202534 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" event={"ID":"8306420e-9e3d-48a6-87b9-af6c61f0d0ba","Type":"ContainerStarted","Data":"0df1f11b6f6ca22fe95cb08ac215a75ef6b758319ee0d0242c465c55916d7ae6"} Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.202758 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.204184 4744 generic.go:334] "Generic (PLEG): container finished" podID="361e3004-e6f9-4f22-a391-ae768a43fa71" containerID="34ef5894affe992d10e99cc365bb964e85e1ce767cf86344c936942fc614a7e2" exitCode=0 Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.205415 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.205459 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" event={"ID":"361e3004-e6f9-4f22-a391-ae768a43fa71","Type":"ContainerDied","Data":"34ef5894affe992d10e99cc365bb964e85e1ce767cf86344c936942fc614a7e2"} Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.205515 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d7d647849-6cdt4" event={"ID":"361e3004-e6f9-4f22-a391-ae768a43fa71","Type":"ContainerDied","Data":"60d7c42d4fef76c22f1f2f6c5091e7fa3d065ed4750145df16a1894a60f5a2cd"} Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.205539 4744 scope.go:117] "RemoveContainer" containerID="34ef5894affe992d10e99cc365bb964e85e1ce767cf86344c936942fc614a7e2" Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.231737 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5f6598d64-fmxvg" podStartSLOduration=14.231700168 podStartE2EDuration="14.231700168s" podCreationTimestamp="2025-10-08 09:29:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:29:14.222556164 +0000 UTC m=+1049.470201403" watchObservedRunningTime="2025-10-08 09:29:14.231700168 +0000 UTC m=+1049.479345407" Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.271886 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6d98c4654d-2xrfx" podStartSLOduration=11.271860342 podStartE2EDuration="11.271860342s" podCreationTimestamp="2025-10-08 09:29:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:29:14.255570557 +0000 UTC m=+1049.503215816" watchObservedRunningTime="2025-10-08 09:29:14.271860342 +0000 UTC m=+1049.519505571" Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.287722 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6b4cb6945c-kjx6m" podStartSLOduration=17.287696065 podStartE2EDuration="17.287696065s" podCreationTimestamp="2025-10-08 09:28:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:29:14.280702349 +0000 UTC m=+1049.528347608" watchObservedRunningTime="2025-10-08 09:29:14.287696065 +0000 UTC m=+1049.535341324" Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.312751 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" podStartSLOduration=14.312734425 podStartE2EDuration="14.312734425s" podCreationTimestamp="2025-10-08 09:29:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:29:14.308504672 +0000 UTC m=+1049.556149911" watchObservedRunningTime="2025-10-08 09:29:14.312734425 +0000 UTC m=+1049.560379664" Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.409693 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d7d647849-6cdt4"] Oct 08 09:29:14 crc kubenswrapper[4744]: I1008 09:29:14.420708 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d7d647849-6cdt4"] Oct 08 09:29:15 crc kubenswrapper[4744]: I1008 09:29:15.216946 4744 generic.go:334] "Generic (PLEG): container finished" podID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerID="94829c949cfcd7db8c18e78499cecb7a0bd2e2cae60407e426135cc749ea730f" exitCode=0 Oct 08 09:29:15 crc kubenswrapper[4744]: I1008 09:29:15.217040 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04474f89-9bb4-4bb7-aef3-66d519650a4b","Type":"ContainerDied","Data":"94829c949cfcd7db8c18e78499cecb7a0bd2e2cae60407e426135cc749ea730f"} Oct 08 09:29:15 crc kubenswrapper[4744]: I1008 09:29:15.218491 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:15 crc kubenswrapper[4744]: I1008 09:29:15.486117 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="361e3004-e6f9-4f22-a391-ae768a43fa71" path="/var/lib/kubelet/pods/361e3004-e6f9-4f22-a391-ae768a43fa71/volumes" Oct 08 09:29:15 crc kubenswrapper[4744]: I1008 09:29:15.779273 4744 scope.go:117] "RemoveContainer" containerID="34ef5894affe992d10e99cc365bb964e85e1ce767cf86344c936942fc614a7e2" Oct 08 09:29:15 crc kubenswrapper[4744]: E1008 09:29:15.780948 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34ef5894affe992d10e99cc365bb964e85e1ce767cf86344c936942fc614a7e2\": container with ID starting with 34ef5894affe992d10e99cc365bb964e85e1ce767cf86344c936942fc614a7e2 not found: ID does not exist" containerID="34ef5894affe992d10e99cc365bb964e85e1ce767cf86344c936942fc614a7e2" Oct 08 09:29:15 crc kubenswrapper[4744]: I1008 09:29:15.781010 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34ef5894affe992d10e99cc365bb964e85e1ce767cf86344c936942fc614a7e2"} err="failed to get container status \"34ef5894affe992d10e99cc365bb964e85e1ce767cf86344c936942fc614a7e2\": rpc error: code = NotFound desc = could not find container \"34ef5894affe992d10e99cc365bb964e85e1ce767cf86344c936942fc614a7e2\": container with ID starting with 34ef5894affe992d10e99cc365bb964e85e1ce767cf86344c936942fc614a7e2 not found: ID does not exist" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.094131 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.222021 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04474f89-9bb4-4bb7-aef3-66d519650a4b-log-httpd\") pod \"04474f89-9bb4-4bb7-aef3-66d519650a4b\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.223109 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-config-data\") pod \"04474f89-9bb4-4bb7-aef3-66d519650a4b\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.223394 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-scripts\") pod \"04474f89-9bb4-4bb7-aef3-66d519650a4b\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.223545 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04474f89-9bb4-4bb7-aef3-66d519650a4b-run-httpd\") pod \"04474f89-9bb4-4bb7-aef3-66d519650a4b\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.223703 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrjgk\" (UniqueName: \"kubernetes.io/projected/04474f89-9bb4-4bb7-aef3-66d519650a4b-kube-api-access-vrjgk\") pod \"04474f89-9bb4-4bb7-aef3-66d519650a4b\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.223812 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-combined-ca-bundle\") pod \"04474f89-9bb4-4bb7-aef3-66d519650a4b\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.224016 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-sg-core-conf-yaml\") pod \"04474f89-9bb4-4bb7-aef3-66d519650a4b\" (UID: \"04474f89-9bb4-4bb7-aef3-66d519650a4b\") " Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.225104 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04474f89-9bb4-4bb7-aef3-66d519650a4b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "04474f89-9bb4-4bb7-aef3-66d519650a4b" (UID: "04474f89-9bb4-4bb7-aef3-66d519650a4b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.225164 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04474f89-9bb4-4bb7-aef3-66d519650a4b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "04474f89-9bb4-4bb7-aef3-66d519650a4b" (UID: "04474f89-9bb4-4bb7-aef3-66d519650a4b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.225894 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04474f89-9bb4-4bb7-aef3-66d519650a4b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.225929 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/04474f89-9bb4-4bb7-aef3-66d519650a4b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.229154 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04474f89-9bb4-4bb7-aef3-66d519650a4b-kube-api-access-vrjgk" (OuterVolumeSpecName: "kube-api-access-vrjgk") pod "04474f89-9bb4-4bb7-aef3-66d519650a4b" (UID: "04474f89-9bb4-4bb7-aef3-66d519650a4b"). InnerVolumeSpecName "kube-api-access-vrjgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.229811 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-scripts" (OuterVolumeSpecName: "scripts") pod "04474f89-9bb4-4bb7-aef3-66d519650a4b" (UID: "04474f89-9bb4-4bb7-aef3-66d519650a4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.257893 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5ccc6f877f-j2zc9" event={"ID":"203b8276-cafa-4b89-a68d-0f9437ee7161","Type":"ContainerStarted","Data":"df18022e9f357c13ab6bda62a95d96dc25e35dcb5f34b2452de0c8580aeaa8eb"} Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.264618 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.264966 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"04474f89-9bb4-4bb7-aef3-66d519650a4b","Type":"ContainerDied","Data":"8966f601de44cde4dbc5b42e98a3693dd7d4c8a7a817c295ae138720de185af4"} Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.265058 4744 scope.go:117] "RemoveContainer" containerID="c4ccd4d1e8dbb1fc765510aedb9d23ab8535bb63e928a42962e2dd1da163e4db" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.279942 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" event={"ID":"4ca73575-280f-43af-8c34-123a1cf83e92","Type":"ContainerStarted","Data":"2507e7644890304a086327aaf114ca55578b1c02ef036c37bc805b05589f09bf"} Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.284837 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "04474f89-9bb4-4bb7-aef3-66d519650a4b" (UID: "04474f89-9bb4-4bb7-aef3-66d519650a4b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.327618 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrjgk\" (UniqueName: \"kubernetes.io/projected/04474f89-9bb4-4bb7-aef3-66d519650a4b-kube-api-access-vrjgk\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.327652 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.327677 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.340150 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-config-data" (OuterVolumeSpecName: "config-data") pod "04474f89-9bb4-4bb7-aef3-66d519650a4b" (UID: "04474f89-9bb4-4bb7-aef3-66d519650a4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.355569 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04474f89-9bb4-4bb7-aef3-66d519650a4b" (UID: "04474f89-9bb4-4bb7-aef3-66d519650a4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.369522 4744 scope.go:117] "RemoveContainer" containerID="94829c949cfcd7db8c18e78499cecb7a0bd2e2cae60407e426135cc749ea730f" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.405790 4744 scope.go:117] "RemoveContainer" containerID="3aba057c1c76a344474b5d9f3b1483ae61f5c2ffce0b161ab7decf06b319e4de" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.431848 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.431880 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04474f89-9bb4-4bb7-aef3-66d519650a4b-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.624833 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.638415 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.660475 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:16 crc kubenswrapper[4744]: E1008 09:29:16.660901 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerName="ceilometer-notification-agent" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.660921 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerName="ceilometer-notification-agent" Oct 08 09:29:16 crc kubenswrapper[4744]: E1008 09:29:16.660938 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerName="sg-core" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.660945 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerName="sg-core" Oct 08 09:29:16 crc kubenswrapper[4744]: E1008 09:29:16.660955 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="361e3004-e6f9-4f22-a391-ae768a43fa71" containerName="init" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.660962 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="361e3004-e6f9-4f22-a391-ae768a43fa71" containerName="init" Oct 08 09:29:16 crc kubenswrapper[4744]: E1008 09:29:16.660972 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerName="ceilometer-central-agent" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.660978 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerName="ceilometer-central-agent" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.661138 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerName="sg-core" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.661159 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="361e3004-e6f9-4f22-a391-ae768a43fa71" containerName="init" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.661169 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerName="ceilometer-central-agent" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.661184 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" containerName="ceilometer-notification-agent" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.662722 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.666955 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.667201 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.698192 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.737471 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.737617 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.737648 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-config-data\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.737669 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-scripts\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.737694 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj2wz\" (UniqueName: \"kubernetes.io/projected/1a85cba9-27c1-4826-ab3e-2d64f36976b0-kube-api-access-qj2wz\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.737722 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a85cba9-27c1-4826-ab3e-2d64f36976b0-run-httpd\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.737784 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a85cba9-27c1-4826-ab3e-2d64f36976b0-log-httpd\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.839979 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a85cba9-27c1-4826-ab3e-2d64f36976b0-log-httpd\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.840084 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.840205 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.840243 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-config-data\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.840276 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-scripts\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.840317 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj2wz\" (UniqueName: \"kubernetes.io/projected/1a85cba9-27c1-4826-ab3e-2d64f36976b0-kube-api-access-qj2wz\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.840672 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a85cba9-27c1-4826-ab3e-2d64f36976b0-log-httpd\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.841141 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a85cba9-27c1-4826-ab3e-2d64f36976b0-run-httpd\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.840843 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a85cba9-27c1-4826-ab3e-2d64f36976b0-run-httpd\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.844933 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.845263 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.847266 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-scripts\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.849469 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-config-data\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.870957 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj2wz\" (UniqueName: \"kubernetes.io/projected/1a85cba9-27c1-4826-ab3e-2d64f36976b0-kube-api-access-qj2wz\") pod \"ceilometer-0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " pod="openstack/ceilometer-0" Oct 08 09:29:16 crc kubenswrapper[4744]: I1008 09:29:16.984123 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:17 crc kubenswrapper[4744]: I1008 09:29:17.296954 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5ccc6f877f-j2zc9" event={"ID":"203b8276-cafa-4b89-a68d-0f9437ee7161","Type":"ContainerStarted","Data":"c838fa3c3fd1742db8399f8fedf836330e8ad3da1250fbd72b58ccf27376122f"} Oct 08 09:29:17 crc kubenswrapper[4744]: I1008 09:29:17.301653 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" event={"ID":"4ca73575-280f-43af-8c34-123a1cf83e92","Type":"ContainerStarted","Data":"f5bd2880eea20935d8032c665e47042e8f0afda6a6c4a57a95b3854550fb3566"} Oct 08 09:29:17 crc kubenswrapper[4744]: I1008 09:29:17.322326 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5ccc6f877f-j2zc9" podStartSLOduration=13.844741205 podStartE2EDuration="17.322311109s" podCreationTimestamp="2025-10-08 09:29:00 +0000 UTC" firstStartedPulling="2025-10-08 09:29:12.356389937 +0000 UTC m=+1047.604035176" lastFinishedPulling="2025-10-08 09:29:15.833959841 +0000 UTC m=+1051.081605080" observedRunningTime="2025-10-08 09:29:17.317827458 +0000 UTC m=+1052.565472697" watchObservedRunningTime="2025-10-08 09:29:17.322311109 +0000 UTC m=+1052.569956348" Oct 08 09:29:17 crc kubenswrapper[4744]: I1008 09:29:17.353855 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-8578895ff6-xg92r" podStartSLOduration=13.987525422000001 podStartE2EDuration="17.353838291s" podCreationTimestamp="2025-10-08 09:29:00 +0000 UTC" firstStartedPulling="2025-10-08 09:29:12.462462603 +0000 UTC m=+1047.710107842" lastFinishedPulling="2025-10-08 09:29:15.828775472 +0000 UTC m=+1051.076420711" observedRunningTime="2025-10-08 09:29:17.348135669 +0000 UTC m=+1052.595780908" watchObservedRunningTime="2025-10-08 09:29:17.353838291 +0000 UTC m=+1052.601483530" Oct 08 09:29:17 crc kubenswrapper[4744]: I1008 09:29:17.464660 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04474f89-9bb4-4bb7-aef3-66d519650a4b" path="/var/lib/kubelet/pods/04474f89-9bb4-4bb7-aef3-66d519650a4b/volumes" Oct 08 09:29:17 crc kubenswrapper[4744]: I1008 09:29:17.513177 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:18 crc kubenswrapper[4744]: I1008 09:29:18.315345 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a85cba9-27c1-4826-ab3e-2d64f36976b0","Type":"ContainerStarted","Data":"5acc853bf0ee41fbd8fbc13cfb31d938aaddb8f4f2e5c37b2f25b3f56fedbad9"} Oct 08 09:29:18 crc kubenswrapper[4744]: I1008 09:29:18.665881 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6c4d7b4588-6wcr4" Oct 08 09:29:19 crc kubenswrapper[4744]: I1008 09:29:19.338742 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a85cba9-27c1-4826-ab3e-2d64f36976b0","Type":"ContainerStarted","Data":"113e7f942f5c47c4d89bf018744b62d2af61e4a5a919711fca506d0743a92973"} Oct 08 09:29:19 crc kubenswrapper[4744]: I1008 09:29:19.339475 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a85cba9-27c1-4826-ab3e-2d64f36976b0","Type":"ContainerStarted","Data":"8b86a0f4beccfbca0dec1469ad796dcdaae8fdf33dd8a14ba7de03ea3a02d202"} Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.351259 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a85cba9-27c1-4826-ab3e-2d64f36976b0","Type":"ContainerStarted","Data":"263b8d626eada0ff070ac79cd4d9c971efd0c9eb54d293c7641c25bb120c38b8"} Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.377493 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.379230 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.383822 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.384298 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-crm8v" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.385597 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.400567 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.442576 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fd3ee7f-6102-426d-b482-a624a7871f42-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6fd3ee7f-6102-426d-b482-a624a7871f42\") " pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.442647 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6fd3ee7f-6102-426d-b482-a624a7871f42-openstack-config-secret\") pod \"openstackclient\" (UID: \"6fd3ee7f-6102-426d-b482-a624a7871f42\") " pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.442695 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6fd3ee7f-6102-426d-b482-a624a7871f42-openstack-config\") pod \"openstackclient\" (UID: \"6fd3ee7f-6102-426d-b482-a624a7871f42\") " pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.442730 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjss9\" (UniqueName: \"kubernetes.io/projected/6fd3ee7f-6102-426d-b482-a624a7871f42-kube-api-access-bjss9\") pod \"openstackclient\" (UID: \"6fd3ee7f-6102-426d-b482-a624a7871f42\") " pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.544558 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fd3ee7f-6102-426d-b482-a624a7871f42-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6fd3ee7f-6102-426d-b482-a624a7871f42\") " pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.545059 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6fd3ee7f-6102-426d-b482-a624a7871f42-openstack-config-secret\") pod \"openstackclient\" (UID: \"6fd3ee7f-6102-426d-b482-a624a7871f42\") " pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.545186 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6fd3ee7f-6102-426d-b482-a624a7871f42-openstack-config\") pod \"openstackclient\" (UID: \"6fd3ee7f-6102-426d-b482-a624a7871f42\") " pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.545288 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjss9\" (UniqueName: \"kubernetes.io/projected/6fd3ee7f-6102-426d-b482-a624a7871f42-kube-api-access-bjss9\") pod \"openstackclient\" (UID: \"6fd3ee7f-6102-426d-b482-a624a7871f42\") " pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.546312 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6fd3ee7f-6102-426d-b482-a624a7871f42-openstack-config\") pod \"openstackclient\" (UID: \"6fd3ee7f-6102-426d-b482-a624a7871f42\") " pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.555538 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fd3ee7f-6102-426d-b482-a624a7871f42-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6fd3ee7f-6102-426d-b482-a624a7871f42\") " pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.569132 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6fd3ee7f-6102-426d-b482-a624a7871f42-openstack-config-secret\") pod \"openstackclient\" (UID: \"6fd3ee7f-6102-426d-b482-a624a7871f42\") " pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.573252 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjss9\" (UniqueName: \"kubernetes.io/projected/6fd3ee7f-6102-426d-b482-a624a7871f42-kube-api-access-bjss9\") pod \"openstackclient\" (UID: \"6fd3ee7f-6102-426d-b482-a624a7871f42\") " pod="openstack/openstackclient" Oct 08 09:29:20 crc kubenswrapper[4744]: I1008 09:29:20.701534 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.076491 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.152408 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-748d7644cf-c74hj"] Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.152773 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" podUID="993f71a0-3367-4a19-b0b0-b7c75b4d7e16" containerName="dnsmasq-dns" containerID="cri-o://9da75d31e16d877276ed2b0a4eee574afad614037f0011bbf412ba3460f32bbd" gracePeriod=10 Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.263738 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.400087 4744 generic.go:334] "Generic (PLEG): container finished" podID="993f71a0-3367-4a19-b0b0-b7c75b4d7e16" containerID="9da75d31e16d877276ed2b0a4eee574afad614037f0011bbf412ba3460f32bbd" exitCode=0 Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.400210 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" event={"ID":"993f71a0-3367-4a19-b0b0-b7c75b4d7e16","Type":"ContainerDied","Data":"9da75d31e16d877276ed2b0a4eee574afad614037f0011bbf412ba3460f32bbd"} Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.401810 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6fd3ee7f-6102-426d-b482-a624a7871f42","Type":"ContainerStarted","Data":"6131915ba1674bb972e24613a57b19e713fa9bcdca5dadc1731ba1c5a138c242"} Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.905602 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.994728 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-dns-svc\") pod \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.995155 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-ovsdbserver-nb\") pod \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.995279 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4zlk\" (UniqueName: \"kubernetes.io/projected/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-kube-api-access-p4zlk\") pod \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.995861 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-ovsdbserver-sb\") pod \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " Oct 08 09:29:21 crc kubenswrapper[4744]: I1008 09:29:21.995962 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-config\") pod \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\" (UID: \"993f71a0-3367-4a19-b0b0-b7c75b4d7e16\") " Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.013536 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-kube-api-access-p4zlk" (OuterVolumeSpecName: "kube-api-access-p4zlk") pod "993f71a0-3367-4a19-b0b0-b7c75b4d7e16" (UID: "993f71a0-3367-4a19-b0b0-b7c75b4d7e16"). InnerVolumeSpecName "kube-api-access-p4zlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.072153 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "993f71a0-3367-4a19-b0b0-b7c75b4d7e16" (UID: "993f71a0-3367-4a19-b0b0-b7c75b4d7e16"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.090848 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "993f71a0-3367-4a19-b0b0-b7c75b4d7e16" (UID: "993f71a0-3367-4a19-b0b0-b7c75b4d7e16"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.097634 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.097660 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.097670 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4zlk\" (UniqueName: \"kubernetes.io/projected/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-kube-api-access-p4zlk\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.102746 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "993f71a0-3367-4a19-b0b0-b7c75b4d7e16" (UID: "993f71a0-3367-4a19-b0b0-b7c75b4d7e16"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.121906 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-config" (OuterVolumeSpecName: "config") pod "993f71a0-3367-4a19-b0b0-b7c75b4d7e16" (UID: "993f71a0-3367-4a19-b0b0-b7c75b4d7e16"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.199056 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.199088 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/993f71a0-3367-4a19-b0b0-b7c75b4d7e16-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.418438 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a85cba9-27c1-4826-ab3e-2d64f36976b0","Type":"ContainerStarted","Data":"c57cd762d457dd360fa2dcb1bde20c981cfe50afab125cab409d84840d35a49b"} Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.419493 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.432084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" event={"ID":"993f71a0-3367-4a19-b0b0-b7c75b4d7e16","Type":"ContainerDied","Data":"28b9fff0f3e701e110d9460cf9192b01141d52f5c9d9ea58a6e47bf8ee139de0"} Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.432146 4744 scope.go:117] "RemoveContainer" containerID="9da75d31e16d877276ed2b0a4eee574afad614037f0011bbf412ba3460f32bbd" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.432306 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-748d7644cf-c74hj" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.545704 4744 scope.go:117] "RemoveContainer" containerID="8176f4ede4b55fe74a47acd764af270597934eb1d8deb1df71deabc9f9175ada" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.564797 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.630804327 podStartE2EDuration="6.564761653s" podCreationTimestamp="2025-10-08 09:29:16 +0000 UTC" firstStartedPulling="2025-10-08 09:29:17.523617609 +0000 UTC m=+1052.771262838" lastFinishedPulling="2025-10-08 09:29:21.457574925 +0000 UTC m=+1056.705220164" observedRunningTime="2025-10-08 09:29:22.448634078 +0000 UTC m=+1057.696279337" watchObservedRunningTime="2025-10-08 09:29:22.564761653 +0000 UTC m=+1057.812406882" Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.596451 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-748d7644cf-c74hj"] Oct 08 09:29:22 crc kubenswrapper[4744]: I1008 09:29:22.620357 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-748d7644cf-c74hj"] Oct 08 09:29:23 crc kubenswrapper[4744]: I1008 09:29:23.467268 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="993f71a0-3367-4a19-b0b0-b7c75b4d7e16" path="/var/lib/kubelet/pods/993f71a0-3367-4a19-b0b0-b7c75b4d7e16/volumes" Oct 08 09:29:23 crc kubenswrapper[4744]: I1008 09:29:23.469575 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:23 crc kubenswrapper[4744]: I1008 09:29:23.669637 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6d98c4654d-2xrfx" podUID="ea9a2551-0abd-4bce-9785-4c8c7e11b02d" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.149:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 09:29:23 crc kubenswrapper[4744]: I1008 09:29:23.670069 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d98c4654d-2xrfx" podUID="ea9a2551-0abd-4bce-9785-4c8c7e11b02d" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.149:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 09:29:24 crc kubenswrapper[4744]: I1008 09:29:24.715329 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-7f466fb8c4-5nw45" podUID="94715e15-b083-4428-95d6-d7e66990af08" containerName="neutron-api" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 08 09:29:24 crc kubenswrapper[4744]: I1008 09:29:24.716332 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-7f466fb8c4-5nw45" podUID="94715e15-b083-4428-95d6-d7e66990af08" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 08 09:29:24 crc kubenswrapper[4744]: I1008 09:29:24.716479 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-7f466fb8c4-5nw45" podUID="94715e15-b083-4428-95d6-d7e66990af08" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 08 09:29:25 crc kubenswrapper[4744]: I1008 09:29:25.312033 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:26 crc kubenswrapper[4744]: I1008 09:29:26.312734 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.014918 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6d98c4654d-2xrfx" Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.102981 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f6598d64-fmxvg"] Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.103910 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.104058 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api-log" containerID="cri-o://0a23d118075d97f8edfe407163951087016299ba68ba0e808fe230d74f207b23" gracePeriod=30 Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.104536 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api" containerID="cri-o://7cd014939d983f7e9915d23c606e20670fd8733ac31eeb6cdffc4d2381684a63" gracePeriod=30 Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.145191 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": EOF" Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.145299 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": EOF" Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.155600 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": EOF" Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.155704 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": EOF" Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.155770 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": EOF" Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.160113 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": EOF" Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.559184 4744 generic.go:334] "Generic (PLEG): container finished" podID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerID="0a23d118075d97f8edfe407163951087016299ba68ba0e808fe230d74f207b23" exitCode=143 Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.559231 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f6598d64-fmxvg" event={"ID":"a50c5d4e-abd2-43bd-9c64-8248b89fb213","Type":"ContainerDied","Data":"0a23d118075d97f8edfe407163951087016299ba68ba0e808fe230d74f207b23"} Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.722612 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-6b4cb6945c-kjx6m" podUID="d7cda27f-1e6a-4637-b647-54f6a8235ce0" containerName="neutron-api" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.723642 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-6b4cb6945c-kjx6m" podUID="d7cda27f-1e6a-4637-b647-54f6a8235ce0" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 08 09:29:27 crc kubenswrapper[4744]: I1008 09:29:27.734327 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-6b4cb6945c-kjx6m" podUID="d7cda27f-1e6a-4637-b647-54f6a8235ce0" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 08 09:29:28 crc kubenswrapper[4744]: I1008 09:29:28.571730 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-q99xn" event={"ID":"f0779920-88cf-4475-9c77-5de26f893b6b","Type":"ContainerStarted","Data":"bcd8e6d09ceff1ef8d99ee2137a4ec92fad6e768959c489f4623be6ecd49c7ce"} Oct 08 09:29:28 crc kubenswrapper[4744]: I1008 09:29:28.602179 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-q99xn" podStartSLOduration=11.059985861 podStartE2EDuration="47.602161319s" podCreationTimestamp="2025-10-08 09:28:41 +0000 UTC" firstStartedPulling="2025-10-08 09:28:49.670388071 +0000 UTC m=+1024.918033310" lastFinishedPulling="2025-10-08 09:29:26.212563529 +0000 UTC m=+1061.460208768" observedRunningTime="2025-10-08 09:29:28.600630578 +0000 UTC m=+1063.848275817" watchObservedRunningTime="2025-10-08 09:29:28.602161319 +0000 UTC m=+1063.849806558" Oct 08 09:29:28 crc kubenswrapper[4744]: I1008 09:29:28.636127 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6d98c4654d-2xrfx" podUID="ea9a2551-0abd-4bce-9785-4c8c7e11b02d" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.149:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 09:29:33 crc kubenswrapper[4744]: I1008 09:29:33.602259 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": read tcp 10.217.0.2:50926->10.217.0.148:9311: read: connection reset by peer" Oct 08 09:29:33 crc kubenswrapper[4744]: I1008 09:29:33.603392 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": read tcp 10.217.0.2:50912->10.217.0.148:9311: read: connection reset by peer" Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.031353 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.032567 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="ceilometer-central-agent" containerID="cri-o://8b86a0f4beccfbca0dec1469ad796dcdaae8fdf33dd8a14ba7de03ea3a02d202" gracePeriod=30 Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.033506 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="proxy-httpd" containerID="cri-o://c57cd762d457dd360fa2dcb1bde20c981cfe50afab125cab409d84840d35a49b" gracePeriod=30 Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.033567 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="sg-core" containerID="cri-o://263b8d626eada0ff070ac79cd4d9c971efd0c9eb54d293c7641c25bb120c38b8" gracePeriod=30 Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.033611 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="ceilometer-notification-agent" containerID="cri-o://113e7f942f5c47c4d89bf018744b62d2af61e4a5a919711fca506d0743a92973" gracePeriod=30 Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.090527 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 08 09:29:34 crc kubenswrapper[4744]: E1008 09:29:34.282760 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a85cba9_27c1_4826_ab3e_2d64f36976b0.slice/crio-263b8d626eada0ff070ac79cd4d9c971efd0c9eb54d293c7641c25bb120c38b8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a85cba9_27c1_4826_ab3e_2d64f36976b0.slice/crio-conmon-c57cd762d457dd360fa2dcb1bde20c981cfe50afab125cab409d84840d35a49b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a85cba9_27c1_4826_ab3e_2d64f36976b0.slice/crio-conmon-263b8d626eada0ff070ac79cd4d9c971efd0c9eb54d293c7641c25bb120c38b8.scope\": RecentStats: unable to find data in memory cache]" Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.645082 4744 generic.go:334] "Generic (PLEG): container finished" podID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerID="7cd014939d983f7e9915d23c606e20670fd8733ac31eeb6cdffc4d2381684a63" exitCode=0 Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.645145 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f6598d64-fmxvg" event={"ID":"a50c5d4e-abd2-43bd-9c64-8248b89fb213","Type":"ContainerDied","Data":"7cd014939d983f7e9915d23c606e20670fd8733ac31eeb6cdffc4d2381684a63"} Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.650398 4744 generic.go:334] "Generic (PLEG): container finished" podID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerID="c57cd762d457dd360fa2dcb1bde20c981cfe50afab125cab409d84840d35a49b" exitCode=0 Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.650421 4744 generic.go:334] "Generic (PLEG): container finished" podID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerID="263b8d626eada0ff070ac79cd4d9c971efd0c9eb54d293c7641c25bb120c38b8" exitCode=2 Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.650430 4744 generic.go:334] "Generic (PLEG): container finished" podID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerID="113e7f942f5c47c4d89bf018744b62d2af61e4a5a919711fca506d0743a92973" exitCode=0 Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.650444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a85cba9-27c1-4826-ab3e-2d64f36976b0","Type":"ContainerDied","Data":"c57cd762d457dd360fa2dcb1bde20c981cfe50afab125cab409d84840d35a49b"} Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.650462 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a85cba9-27c1-4826-ab3e-2d64f36976b0","Type":"ContainerDied","Data":"263b8d626eada0ff070ac79cd4d9c971efd0c9eb54d293c7641c25bb120c38b8"} Oct 08 09:29:34 crc kubenswrapper[4744]: I1008 09:29:34.650474 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a85cba9-27c1-4826-ab3e-2d64f36976b0","Type":"ContainerDied","Data":"113e7f942f5c47c4d89bf018744b62d2af61e4a5a919711fca506d0743a92973"} Oct 08 09:29:35 crc kubenswrapper[4744]: I1008 09:29:35.673281 4744 generic.go:334] "Generic (PLEG): container finished" podID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerID="8b86a0f4beccfbca0dec1469ad796dcdaae8fdf33dd8a14ba7de03ea3a02d202" exitCode=0 Oct 08 09:29:35 crc kubenswrapper[4744]: I1008 09:29:35.673454 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a85cba9-27c1-4826-ab3e-2d64f36976b0","Type":"ContainerDied","Data":"8b86a0f4beccfbca0dec1469ad796dcdaae8fdf33dd8a14ba7de03ea3a02d202"} Oct 08 09:29:35 crc kubenswrapper[4744]: I1008 09:29:35.676769 4744 generic.go:334] "Generic (PLEG): container finished" podID="f0779920-88cf-4475-9c77-5de26f893b6b" containerID="bcd8e6d09ceff1ef8d99ee2137a4ec92fad6e768959c489f4623be6ecd49c7ce" exitCode=0 Oct 08 09:29:35 crc kubenswrapper[4744]: I1008 09:29:35.676826 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-q99xn" event={"ID":"f0779920-88cf-4475-9c77-5de26f893b6b","Type":"ContainerDied","Data":"bcd8e6d09ceff1ef8d99ee2137a4ec92fad6e768959c489f4623be6ecd49c7ce"} Oct 08 09:29:36 crc kubenswrapper[4744]: I1008 09:29:36.273561 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": dial tcp 10.217.0.148:9311: connect: connection refused" Oct 08 09:29:36 crc kubenswrapper[4744]: I1008 09:29:36.273643 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f6598d64-fmxvg" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.148:9311/healthcheck\": dial tcp 10.217.0.148:9311: connect: connection refused" Oct 08 09:29:36 crc kubenswrapper[4744]: I1008 09:29:36.273688 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.275730 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.300124 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.397590 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-q99xn" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.424213 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a50c5d4e-abd2-43bd-9c64-8248b89fb213-logs\") pod \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.424284 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-config-data-custom\") pod \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.424424 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qj2wz\" (UniqueName: \"kubernetes.io/projected/1a85cba9-27c1-4826-ab3e-2d64f36976b0-kube-api-access-qj2wz\") pod \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.424474 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-scripts\") pod \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.424500 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-combined-ca-bundle\") pod \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.424627 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-config-data\") pod \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.424647 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-sg-core-conf-yaml\") pod \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.424702 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-config-data\") pod \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.424738 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vslb9\" (UniqueName: \"kubernetes.io/projected/a50c5d4e-abd2-43bd-9c64-8248b89fb213-kube-api-access-vslb9\") pod \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\" (UID: \"a50c5d4e-abd2-43bd-9c64-8248b89fb213\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.424816 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a85cba9-27c1-4826-ab3e-2d64f36976b0-run-httpd\") pod \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.424838 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-combined-ca-bundle\") pod \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.424867 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a85cba9-27c1-4826-ab3e-2d64f36976b0-log-httpd\") pod \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\" (UID: \"1a85cba9-27c1-4826-ab3e-2d64f36976b0\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.425481 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a50c5d4e-abd2-43bd-9c64-8248b89fb213-logs" (OuterVolumeSpecName: "logs") pod "a50c5d4e-abd2-43bd-9c64-8248b89fb213" (UID: "a50c5d4e-abd2-43bd-9c64-8248b89fb213"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.425847 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a85cba9-27c1-4826-ab3e-2d64f36976b0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1a85cba9-27c1-4826-ab3e-2d64f36976b0" (UID: "1a85cba9-27c1-4826-ab3e-2d64f36976b0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.428644 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a85cba9-27c1-4826-ab3e-2d64f36976b0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1a85cba9-27c1-4826-ab3e-2d64f36976b0" (UID: "1a85cba9-27c1-4826-ab3e-2d64f36976b0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.462560 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a50c5d4e-abd2-43bd-9c64-8248b89fb213" (UID: "a50c5d4e-abd2-43bd-9c64-8248b89fb213"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.467043 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-scripts" (OuterVolumeSpecName: "scripts") pod "1a85cba9-27c1-4826-ab3e-2d64f36976b0" (UID: "1a85cba9-27c1-4826-ab3e-2d64f36976b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.467793 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a85cba9-27c1-4826-ab3e-2d64f36976b0-kube-api-access-qj2wz" (OuterVolumeSpecName: "kube-api-access-qj2wz") pod "1a85cba9-27c1-4826-ab3e-2d64f36976b0" (UID: "1a85cba9-27c1-4826-ab3e-2d64f36976b0"). InnerVolumeSpecName "kube-api-access-qj2wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.487521 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a50c5d4e-abd2-43bd-9c64-8248b89fb213-kube-api-access-vslb9" (OuterVolumeSpecName: "kube-api-access-vslb9") pod "a50c5d4e-abd2-43bd-9c64-8248b89fb213" (UID: "a50c5d4e-abd2-43bd-9c64-8248b89fb213"). InnerVolumeSpecName "kube-api-access-vslb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.527175 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-db-sync-config-data\") pod \"f0779920-88cf-4475-9c77-5de26f893b6b\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.527302 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-scripts\") pod \"f0779920-88cf-4475-9c77-5de26f893b6b\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.527397 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-config-data\") pod \"f0779920-88cf-4475-9c77-5de26f893b6b\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.527417 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stp2z\" (UniqueName: \"kubernetes.io/projected/f0779920-88cf-4475-9c77-5de26f893b6b-kube-api-access-stp2z\") pod \"f0779920-88cf-4475-9c77-5de26f893b6b\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.527514 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0779920-88cf-4475-9c77-5de26f893b6b-etc-machine-id\") pod \"f0779920-88cf-4475-9c77-5de26f893b6b\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.527600 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-combined-ca-bundle\") pod \"f0779920-88cf-4475-9c77-5de26f893b6b\" (UID: \"f0779920-88cf-4475-9c77-5de26f893b6b\") " Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.528069 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vslb9\" (UniqueName: \"kubernetes.io/projected/a50c5d4e-abd2-43bd-9c64-8248b89fb213-kube-api-access-vslb9\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.528088 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a85cba9-27c1-4826-ab3e-2d64f36976b0-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.528097 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a85cba9-27c1-4826-ab3e-2d64f36976b0-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.528106 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a50c5d4e-abd2-43bd-9c64-8248b89fb213-logs\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.528115 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.528124 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qj2wz\" (UniqueName: \"kubernetes.io/projected/1a85cba9-27c1-4826-ab3e-2d64f36976b0-kube-api-access-qj2wz\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.528134 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.539267 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0779920-88cf-4475-9c77-5de26f893b6b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f0779920-88cf-4475-9c77-5de26f893b6b" (UID: "f0779920-88cf-4475-9c77-5de26f893b6b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.547497 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0779920-88cf-4475-9c77-5de26f893b6b-kube-api-access-stp2z" (OuterVolumeSpecName: "kube-api-access-stp2z") pod "f0779920-88cf-4475-9c77-5de26f893b6b" (UID: "f0779920-88cf-4475-9c77-5de26f893b6b"). InnerVolumeSpecName "kube-api-access-stp2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.548834 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f0779920-88cf-4475-9c77-5de26f893b6b" (UID: "f0779920-88cf-4475-9c77-5de26f893b6b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.566993 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-scripts" (OuterVolumeSpecName: "scripts") pod "f0779920-88cf-4475-9c77-5de26f893b6b" (UID: "f0779920-88cf-4475-9c77-5de26f893b6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.571471 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1a85cba9-27c1-4826-ab3e-2d64f36976b0" (UID: "1a85cba9-27c1-4826-ab3e-2d64f36976b0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.599777 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a50c5d4e-abd2-43bd-9c64-8248b89fb213" (UID: "a50c5d4e-abd2-43bd-9c64-8248b89fb213"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.599962 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0779920-88cf-4475-9c77-5de26f893b6b" (UID: "f0779920-88cf-4475-9c77-5de26f893b6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.622189 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-config-data" (OuterVolumeSpecName: "config-data") pod "a50c5d4e-abd2-43bd-9c64-8248b89fb213" (UID: "a50c5d4e-abd2-43bd-9c64-8248b89fb213"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.630775 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stp2z\" (UniqueName: \"kubernetes.io/projected/f0779920-88cf-4475-9c77-5de26f893b6b-kube-api-access-stp2z\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.630812 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.630825 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0779920-88cf-4475-9c77-5de26f893b6b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.630838 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a50c5d4e-abd2-43bd-9c64-8248b89fb213-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.630850 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.630861 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.630875 4744 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.630885 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.672022 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a85cba9-27c1-4826-ab3e-2d64f36976b0" (UID: "1a85cba9-27c1-4826-ab3e-2d64f36976b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.674655 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-config-data" (OuterVolumeSpecName: "config-data") pod "f0779920-88cf-4475-9c77-5de26f893b6b" (UID: "f0779920-88cf-4475-9c77-5de26f893b6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.681694 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-config-data" (OuterVolumeSpecName: "config-data") pod "1a85cba9-27c1-4826-ab3e-2d64f36976b0" (UID: "1a85cba9-27c1-4826-ab3e-2d64f36976b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.698346 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6fd3ee7f-6102-426d-b482-a624a7871f42","Type":"ContainerStarted","Data":"871b3971530a72c24fb146bb0ddabbac22de83d0787d9c8bb4fa32a6aad7ea1c"} Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.702933 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a85cba9-27c1-4826-ab3e-2d64f36976b0","Type":"ContainerDied","Data":"5acc853bf0ee41fbd8fbc13cfb31d938aaddb8f4f2e5c37b2f25b3f56fedbad9"} Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.702973 4744 scope.go:117] "RemoveContainer" containerID="c57cd762d457dd360fa2dcb1bde20c981cfe50afab125cab409d84840d35a49b" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.703141 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.708674 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-q99xn" event={"ID":"f0779920-88cf-4475-9c77-5de26f893b6b","Type":"ContainerDied","Data":"74af59037c0762522316ca67cae1475bf2d618fd155f9f3f801b215101498614"} Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.708705 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74af59037c0762522316ca67cae1475bf2d618fd155f9f3f801b215101498614" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.708756 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-q99xn" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.736082 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f6598d64-fmxvg" event={"ID":"a50c5d4e-abd2-43bd-9c64-8248b89fb213","Type":"ContainerDied","Data":"d03420725a90694970cefe57d1af847a35d5d15a69b61b15f5410e201b065212"} Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.736283 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f6598d64-fmxvg" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.738413 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.738441 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a85cba9-27c1-4826-ab3e-2d64f36976b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.738453 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0779920-88cf-4475-9c77-5de26f893b6b-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.744422 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.246828743 podStartE2EDuration="17.744399844s" podCreationTimestamp="2025-10-08 09:29:20 +0000 UTC" firstStartedPulling="2025-10-08 09:29:21.273526384 +0000 UTC m=+1056.521171623" lastFinishedPulling="2025-10-08 09:29:36.771097485 +0000 UTC m=+1072.018742724" observedRunningTime="2025-10-08 09:29:37.725995752 +0000 UTC m=+1072.973640991" watchObservedRunningTime="2025-10-08 09:29:37.744399844 +0000 UTC m=+1072.992045083" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.784859 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.787638 4744 scope.go:117] "RemoveContainer" containerID="263b8d626eada0ff070ac79cd4d9c971efd0c9eb54d293c7641c25bb120c38b8" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.792601 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.815710 4744 scope.go:117] "RemoveContainer" containerID="113e7f942f5c47c4d89bf018744b62d2af61e4a5a919711fca506d0743a92973" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.827128 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:37 crc kubenswrapper[4744]: E1008 09:29:37.827740 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="993f71a0-3367-4a19-b0b0-b7c75b4d7e16" containerName="dnsmasq-dns" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.827764 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="993f71a0-3367-4a19-b0b0-b7c75b4d7e16" containerName="dnsmasq-dns" Oct 08 09:29:37 crc kubenswrapper[4744]: E1008 09:29:37.827786 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="proxy-httpd" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.827794 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="proxy-httpd" Oct 08 09:29:37 crc kubenswrapper[4744]: E1008 09:29:37.827811 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="sg-core" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.827818 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="sg-core" Oct 08 09:29:37 crc kubenswrapper[4744]: E1008 09:29:37.827836 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="993f71a0-3367-4a19-b0b0-b7c75b4d7e16" containerName="init" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.827844 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="993f71a0-3367-4a19-b0b0-b7c75b4d7e16" containerName="init" Oct 08 09:29:37 crc kubenswrapper[4744]: E1008 09:29:37.827856 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0779920-88cf-4475-9c77-5de26f893b6b" containerName="cinder-db-sync" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.827863 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0779920-88cf-4475-9c77-5de26f893b6b" containerName="cinder-db-sync" Oct 08 09:29:37 crc kubenswrapper[4744]: E1008 09:29:37.827879 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="ceilometer-central-agent" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.827885 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="ceilometer-central-agent" Oct 08 09:29:37 crc kubenswrapper[4744]: E1008 09:29:37.827897 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="ceilometer-notification-agent" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.827902 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="ceilometer-notification-agent" Oct 08 09:29:37 crc kubenswrapper[4744]: E1008 09:29:37.827919 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api-log" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.827926 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api-log" Oct 08 09:29:37 crc kubenswrapper[4744]: E1008 09:29:37.827938 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.827945 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.828125 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="ceilometer-central-agent" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.828137 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="993f71a0-3367-4a19-b0b0-b7c75b4d7e16" containerName="dnsmasq-dns" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.828151 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.828160 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0779920-88cf-4475-9c77-5de26f893b6b" containerName="cinder-db-sync" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.828171 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" containerName="barbican-api-log" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.828181 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="ceilometer-notification-agent" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.828189 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="sg-core" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.828202 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" containerName="proxy-httpd" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.830171 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.841064 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.841290 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.850569 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f6598d64-fmxvg"] Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.873626 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5f6598d64-fmxvg"] Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.876754 4744 scope.go:117] "RemoveContainer" containerID="8b86a0f4beccfbca0dec1469ad796dcdaae8fdf33dd8a14ba7de03ea3a02d202" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.892442 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.916272 4744 scope.go:117] "RemoveContainer" containerID="7cd014939d983f7e9915d23c606e20670fd8733ac31eeb6cdffc4d2381684a63" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.955533 4744 scope.go:117] "RemoveContainer" containerID="0a23d118075d97f8edfe407163951087016299ba68ba0e808fe230d74f207b23" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.957034 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8536785-c783-4780-9135-4a460f274c68-run-httpd\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.957097 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.957145 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-config-data\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.957191 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqz57\" (UniqueName: \"kubernetes.io/projected/a8536785-c783-4780-9135-4a460f274c68-kube-api-access-fqz57\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.957233 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.957531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-scripts\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:37 crc kubenswrapper[4744]: I1008 09:29:37.957735 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8536785-c783-4780-9135-4a460f274c68-log-httpd\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.059949 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-config-data\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.060490 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqz57\" (UniqueName: \"kubernetes.io/projected/a8536785-c783-4780-9135-4a460f274c68-kube-api-access-fqz57\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.060538 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.060595 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-scripts\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.060646 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8536785-c783-4780-9135-4a460f274c68-log-httpd\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.060665 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8536785-c783-4780-9135-4a460f274c68-run-httpd\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.060707 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.062233 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8536785-c783-4780-9135-4a460f274c68-run-httpd\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.062506 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8536785-c783-4780-9135-4a460f274c68-log-httpd\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.068101 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.068740 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.072395 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-scripts\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.077489 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-config-data\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.103746 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.105896 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.134173 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.134447 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.134637 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.134847 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-q2q4j" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.144127 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqz57\" (UniqueName: \"kubernetes.io/projected/a8536785-c783-4780-9135-4a460f274c68-kube-api-access-fqz57\") pod \"ceilometer-0\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.173489 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.178582 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7"] Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.180991 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.268229 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7"] Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.269429 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-scripts\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.269480 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.269542 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a257281-2eda-4f01-920f-02757ef0fb37-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.269588 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.269619 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhtsp\" (UniqueName: \"kubernetes.io/projected/8a257281-2eda-4f01-920f-02757ef0fb37-kube-api-access-lhtsp\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.269658 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-config-data\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.301314 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.373089 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.373134 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhtsp\" (UniqueName: \"kubernetes.io/projected/8a257281-2eda-4f01-920f-02757ef0fb37-kube-api-access-lhtsp\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.373167 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-dns-svc\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.373195 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7ctw\" (UniqueName: \"kubernetes.io/projected/24751171-e89c-41e2-b441-654ba81d749b-kube-api-access-d7ctw\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.373219 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-ovsdbserver-nb\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.373241 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-config-data\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.373269 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-config\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.373331 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-scripts\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.373353 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.373397 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-ovsdbserver-sb\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.373441 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a257281-2eda-4f01-920f-02757ef0fb37-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.373526 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a257281-2eda-4f01-920f-02757ef0fb37-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.390708 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-config-data\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.391085 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-scripts\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.410577 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.416450 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.416888 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.421694 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.437812 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhtsp\" (UniqueName: \"kubernetes.io/projected/8a257281-2eda-4f01-920f-02757ef0fb37-kube-api-access-lhtsp\") pod \"cinder-scheduler-0\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.438295 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.448495 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.504214 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.528872 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-config-data-custom\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.528970 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1126d25d-8501-4fd5-9ac6-cde91e58ac59-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.529002 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-config-data\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.529033 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-dns-svc\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.529099 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7ctw\" (UniqueName: \"kubernetes.io/projected/24751171-e89c-41e2-b441-654ba81d749b-kube-api-access-d7ctw\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.529134 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-ovsdbserver-nb\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.529160 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1126d25d-8501-4fd5-9ac6-cde91e58ac59-logs\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.529222 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-config\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.529278 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.536028 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhvr7\" (UniqueName: \"kubernetes.io/projected/1126d25d-8501-4fd5-9ac6-cde91e58ac59-kube-api-access-nhvr7\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.536138 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-scripts\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.536177 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-ovsdbserver-sb\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.539527 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-ovsdbserver-sb\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.546296 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-ovsdbserver-nb\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.548093 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-dns-svc\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.577792 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-config\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.585003 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7ctw\" (UniqueName: \"kubernetes.io/projected/24751171-e89c-41e2-b441-654ba81d749b-kube-api-access-d7ctw\") pod \"dnsmasq-dns-7bdc9d6cdc-8rhc7\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.653576 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.653676 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhvr7\" (UniqueName: \"kubernetes.io/projected/1126d25d-8501-4fd5-9ac6-cde91e58ac59-kube-api-access-nhvr7\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.653715 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-scripts\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.653767 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-config-data-custom\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.653802 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1126d25d-8501-4fd5-9ac6-cde91e58ac59-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.653824 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-config-data\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.653856 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1126d25d-8501-4fd5-9ac6-cde91e58ac59-logs\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.659743 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1126d25d-8501-4fd5-9ac6-cde91e58ac59-logs\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.662172 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1126d25d-8501-4fd5-9ac6-cde91e58ac59-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.663251 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-scripts\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.681791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-config-data-custom\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.683407 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.734653 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-config-data\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.762176 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhvr7\" (UniqueName: \"kubernetes.io/projected/1126d25d-8501-4fd5-9ac6-cde91e58ac59-kube-api-access-nhvr7\") pod \"cinder-api-0\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.769469 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-cmhgp"] Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.770892 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cmhgp" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.805090 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.807107 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.869111 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slwrl\" (UniqueName: \"kubernetes.io/projected/56496cc2-c5e8-426d-8f7c-fbd84f7b266c-kube-api-access-slwrl\") pod \"nova-api-db-create-cmhgp\" (UID: \"56496cc2-c5e8-426d-8f7c-fbd84f7b266c\") " pod="openstack/nova-api-db-create-cmhgp" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.873254 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-cmhgp"] Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.946245 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-58rvm"] Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.951491 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-58rvm" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.982140 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slwrl\" (UniqueName: \"kubernetes.io/projected/56496cc2-c5e8-426d-8f7c-fbd84f7b266c-kube-api-access-slwrl\") pod \"nova-api-db-create-cmhgp\" (UID: \"56496cc2-c5e8-426d-8f7c-fbd84f7b266c\") " pod="openstack/nova-api-db-create-cmhgp" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.982709 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crgrr\" (UniqueName: \"kubernetes.io/projected/daa62f04-afb7-4d1e-9218-cbc2cc22e807-kube-api-access-crgrr\") pod \"nova-cell0-db-create-58rvm\" (UID: \"daa62f04-afb7-4d1e-9218-cbc2cc22e807\") " pod="openstack/nova-cell0-db-create-58rvm" Oct 08 09:29:38 crc kubenswrapper[4744]: I1008 09:29:38.994763 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-58rvm"] Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.017083 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slwrl\" (UniqueName: \"kubernetes.io/projected/56496cc2-c5e8-426d-8f7c-fbd84f7b266c-kube-api-access-slwrl\") pod \"nova-api-db-create-cmhgp\" (UID: \"56496cc2-c5e8-426d-8f7c-fbd84f7b266c\") " pod="openstack/nova-api-db-create-cmhgp" Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.058332 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.066824 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-plnkl"] Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.074223 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-plnkl" Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.084134 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crgrr\" (UniqueName: \"kubernetes.io/projected/daa62f04-afb7-4d1e-9218-cbc2cc22e807-kube-api-access-crgrr\") pod \"nova-cell0-db-create-58rvm\" (UID: \"daa62f04-afb7-4d1e-9218-cbc2cc22e807\") " pod="openstack/nova-cell0-db-create-58rvm" Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.120568 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-plnkl"] Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.121287 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crgrr\" (UniqueName: \"kubernetes.io/projected/daa62f04-afb7-4d1e-9218-cbc2cc22e807-kube-api-access-crgrr\") pod \"nova-cell0-db-create-58rvm\" (UID: \"daa62f04-afb7-4d1e-9218-cbc2cc22e807\") " pod="openstack/nova-cell0-db-create-58rvm" Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.150341 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cmhgp" Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.186537 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78khw\" (UniqueName: \"kubernetes.io/projected/929c71a8-07ed-475b-b150-e5a1d92e22c5-kube-api-access-78khw\") pod \"nova-cell1-db-create-plnkl\" (UID: \"929c71a8-07ed-475b-b150-e5a1d92e22c5\") " pod="openstack/nova-cell1-db-create-plnkl" Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.296861 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78khw\" (UniqueName: \"kubernetes.io/projected/929c71a8-07ed-475b-b150-e5a1d92e22c5-kube-api-access-78khw\") pod \"nova-cell1-db-create-plnkl\" (UID: \"929c71a8-07ed-475b-b150-e5a1d92e22c5\") " pod="openstack/nova-cell1-db-create-plnkl" Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.298178 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-58rvm" Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.323190 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78khw\" (UniqueName: \"kubernetes.io/projected/929c71a8-07ed-475b-b150-e5a1d92e22c5-kube-api-access-78khw\") pod \"nova-cell1-db-create-plnkl\" (UID: \"929c71a8-07ed-475b-b150-e5a1d92e22c5\") " pod="openstack/nova-cell1-db-create-plnkl" Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.431501 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-plnkl" Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.486128 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a85cba9-27c1-4826-ab3e-2d64f36976b0" path="/var/lib/kubelet/pods/1a85cba9-27c1-4826-ab3e-2d64f36976b0/volumes" Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.487248 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a50c5d4e-abd2-43bd-9c64-8248b89fb213" path="/var/lib/kubelet/pods/a50c5d4e-abd2-43bd-9c64-8248b89fb213/volumes" Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.584477 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.658860 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.678869 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7"] Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.903018 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-cmhgp"] Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.928519 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8a257281-2eda-4f01-920f-02757ef0fb37","Type":"ContainerStarted","Data":"8159019b86efb3037ce6bc5db6141800383b41c1982f22c60b4fa072bfb01d13"} Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.930845 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1126d25d-8501-4fd5-9ac6-cde91e58ac59","Type":"ContainerStarted","Data":"8ed9ad67eaa32d8ae1bca1dd2c5f2946008c10628211b7ceb47aa68f2fa15d7a"} Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.932919 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" event={"ID":"24751171-e89c-41e2-b441-654ba81d749b","Type":"ContainerStarted","Data":"ca6664161f914fa99ff1e59491979bd7130a74853e8a16f9668cb0b0e7de592b"} Oct 08 09:29:39 crc kubenswrapper[4744]: I1008 09:29:39.944328 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8536785-c783-4780-9135-4a460f274c68","Type":"ContainerStarted","Data":"09e3cdc99ecfcdaab4e8ae30f37b0982c0d1a73274e978e0b59df8f036431f08"} Oct 08 09:29:40 crc kubenswrapper[4744]: I1008 09:29:40.151225 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-58rvm"] Oct 08 09:29:40 crc kubenswrapper[4744]: W1008 09:29:40.177563 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaa62f04_afb7_4d1e_9218_cbc2cc22e807.slice/crio-5e2caef7fbe2e8f4bb44de6758dd14a93722098d35fc028a5e245e09d9874da5 WatchSource:0}: Error finding container 5e2caef7fbe2e8f4bb44de6758dd14a93722098d35fc028a5e245e09d9874da5: Status 404 returned error can't find the container with id 5e2caef7fbe2e8f4bb44de6758dd14a93722098d35fc028a5e245e09d9874da5 Oct 08 09:29:40 crc kubenswrapper[4744]: I1008 09:29:40.424607 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-plnkl"] Oct 08 09:29:40 crc kubenswrapper[4744]: W1008 09:29:40.457781 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod929c71a8_07ed_475b_b150_e5a1d92e22c5.slice/crio-1098ce5d3939831583b980dfaa72dd6f675c376418ca6dc821fa77d31e2b1417 WatchSource:0}: Error finding container 1098ce5d3939831583b980dfaa72dd6f675c376418ca6dc821fa77d31e2b1417: Status 404 returned error can't find the container with id 1098ce5d3939831583b980dfaa72dd6f675c376418ca6dc821fa77d31e2b1417 Oct 08 09:29:40 crc kubenswrapper[4744]: I1008 09:29:40.975767 4744 generic.go:334] "Generic (PLEG): container finished" podID="daa62f04-afb7-4d1e-9218-cbc2cc22e807" containerID="3361f212ebd338ebbe7433cc11d50c31512677ad8628093354c6f994e160c388" exitCode=0 Oct 08 09:29:40 crc kubenswrapper[4744]: I1008 09:29:40.975859 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-58rvm" event={"ID":"daa62f04-afb7-4d1e-9218-cbc2cc22e807","Type":"ContainerDied","Data":"3361f212ebd338ebbe7433cc11d50c31512677ad8628093354c6f994e160c388"} Oct 08 09:29:40 crc kubenswrapper[4744]: I1008 09:29:40.976324 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-58rvm" event={"ID":"daa62f04-afb7-4d1e-9218-cbc2cc22e807","Type":"ContainerStarted","Data":"5e2caef7fbe2e8f4bb44de6758dd14a93722098d35fc028a5e245e09d9874da5"} Oct 08 09:29:40 crc kubenswrapper[4744]: I1008 09:29:40.996706 4744 generic.go:334] "Generic (PLEG): container finished" podID="24751171-e89c-41e2-b441-654ba81d749b" containerID="322089b1ad280f38bee2713c7ce87b5a3d854fd301692db2a2f307edfc788b52" exitCode=0 Oct 08 09:29:40 crc kubenswrapper[4744]: I1008 09:29:40.996773 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" event={"ID":"24751171-e89c-41e2-b441-654ba81d749b","Type":"ContainerDied","Data":"322089b1ad280f38bee2713c7ce87b5a3d854fd301692db2a2f307edfc788b52"} Oct 08 09:29:41 crc kubenswrapper[4744]: I1008 09:29:41.019648 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8536785-c783-4780-9135-4a460f274c68","Type":"ContainerStarted","Data":"f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5"} Oct 08 09:29:41 crc kubenswrapper[4744]: I1008 09:29:41.049767 4744 generic.go:334] "Generic (PLEG): container finished" podID="56496cc2-c5e8-426d-8f7c-fbd84f7b266c" containerID="452f48ca078c410657ab748a065813649452a099e3162cc9e97b35a02c5b1efc" exitCode=0 Oct 08 09:29:41 crc kubenswrapper[4744]: I1008 09:29:41.049843 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cmhgp" event={"ID":"56496cc2-c5e8-426d-8f7c-fbd84f7b266c","Type":"ContainerDied","Data":"452f48ca078c410657ab748a065813649452a099e3162cc9e97b35a02c5b1efc"} Oct 08 09:29:41 crc kubenswrapper[4744]: I1008 09:29:41.049880 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cmhgp" event={"ID":"56496cc2-c5e8-426d-8f7c-fbd84f7b266c","Type":"ContainerStarted","Data":"ff318c70caa76a0ba1610d4b3cdbd8dc0e69974a00ead06dd4e961f52bbdc6eb"} Oct 08 09:29:41 crc kubenswrapper[4744]: I1008 09:29:41.057969 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-plnkl" event={"ID":"929c71a8-07ed-475b-b150-e5a1d92e22c5","Type":"ContainerStarted","Data":"a817ffeda72b2f03ed83fd5396bca4ad58202d5e6b39111d24b5e40c515ea1ad"} Oct 08 09:29:41 crc kubenswrapper[4744]: I1008 09:29:41.058019 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-plnkl" event={"ID":"929c71a8-07ed-475b-b150-e5a1d92e22c5","Type":"ContainerStarted","Data":"1098ce5d3939831583b980dfaa72dd6f675c376418ca6dc821fa77d31e2b1417"} Oct 08 09:29:41 crc kubenswrapper[4744]: I1008 09:29:41.231154 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-plnkl" podStartSLOduration=2.231122723 podStartE2EDuration="2.231122723s" podCreationTimestamp="2025-10-08 09:29:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:29:41.128356136 +0000 UTC m=+1076.376001375" watchObservedRunningTime="2025-10-08 09:29:41.231122723 +0000 UTC m=+1076.478767962" Oct 08 09:29:41 crc kubenswrapper[4744]: I1008 09:29:41.567561 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.081544 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8536785-c783-4780-9135-4a460f274c68","Type":"ContainerStarted","Data":"6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d"} Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.083873 4744 generic.go:334] "Generic (PLEG): container finished" podID="929c71a8-07ed-475b-b150-e5a1d92e22c5" containerID="a817ffeda72b2f03ed83fd5396bca4ad58202d5e6b39111d24b5e40c515ea1ad" exitCode=0 Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.083949 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-plnkl" event={"ID":"929c71a8-07ed-475b-b150-e5a1d92e22c5","Type":"ContainerDied","Data":"a817ffeda72b2f03ed83fd5396bca4ad58202d5e6b39111d24b5e40c515ea1ad"} Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.086186 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1126d25d-8501-4fd5-9ac6-cde91e58ac59","Type":"ContainerStarted","Data":"0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e"} Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.094172 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" event={"ID":"24751171-e89c-41e2-b441-654ba81d749b","Type":"ContainerStarted","Data":"9bc3f464205ea70d2f6d5ae1f40e9c86a7bfc698e87ec37cdfb100842802c36f"} Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.094777 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.185678 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" podStartSLOduration=4.18565273 podStartE2EDuration="4.18565273s" podCreationTimestamp="2025-10-08 09:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:29:42.182440094 +0000 UTC m=+1077.430085333" watchObservedRunningTime="2025-10-08 09:29:42.18565273 +0000 UTC m=+1077.433297969" Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.815676 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-58rvm" Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.818740 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cmhgp" Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.944518 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slwrl\" (UniqueName: \"kubernetes.io/projected/56496cc2-c5e8-426d-8f7c-fbd84f7b266c-kube-api-access-slwrl\") pod \"56496cc2-c5e8-426d-8f7c-fbd84f7b266c\" (UID: \"56496cc2-c5e8-426d-8f7c-fbd84f7b266c\") " Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.944599 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crgrr\" (UniqueName: \"kubernetes.io/projected/daa62f04-afb7-4d1e-9218-cbc2cc22e807-kube-api-access-crgrr\") pod \"daa62f04-afb7-4d1e-9218-cbc2cc22e807\" (UID: \"daa62f04-afb7-4d1e-9218-cbc2cc22e807\") " Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.986313 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56496cc2-c5e8-426d-8f7c-fbd84f7b266c-kube-api-access-slwrl" (OuterVolumeSpecName: "kube-api-access-slwrl") pod "56496cc2-c5e8-426d-8f7c-fbd84f7b266c" (UID: "56496cc2-c5e8-426d-8f7c-fbd84f7b266c"). InnerVolumeSpecName "kube-api-access-slwrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:42 crc kubenswrapper[4744]: I1008 09:29:42.987165 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daa62f04-afb7-4d1e-9218-cbc2cc22e807-kube-api-access-crgrr" (OuterVolumeSpecName: "kube-api-access-crgrr") pod "daa62f04-afb7-4d1e-9218-cbc2cc22e807" (UID: "daa62f04-afb7-4d1e-9218-cbc2cc22e807"). InnerVolumeSpecName "kube-api-access-crgrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.047617 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slwrl\" (UniqueName: \"kubernetes.io/projected/56496cc2-c5e8-426d-8f7c-fbd84f7b266c-kube-api-access-slwrl\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.047661 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crgrr\" (UniqueName: \"kubernetes.io/projected/daa62f04-afb7-4d1e-9218-cbc2cc22e807-kube-api-access-crgrr\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.110200 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8a257281-2eda-4f01-920f-02757ef0fb37","Type":"ContainerStarted","Data":"3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880"} Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.111847 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cmhgp" event={"ID":"56496cc2-c5e8-426d-8f7c-fbd84f7b266c","Type":"ContainerDied","Data":"ff318c70caa76a0ba1610d4b3cdbd8dc0e69974a00ead06dd4e961f52bbdc6eb"} Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.111875 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff318c70caa76a0ba1610d4b3cdbd8dc0e69974a00ead06dd4e961f52bbdc6eb" Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.111943 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cmhgp" Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.152665 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-58rvm" Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.153141 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-58rvm" event={"ID":"daa62f04-afb7-4d1e-9218-cbc2cc22e807","Type":"ContainerDied","Data":"5e2caef7fbe2e8f4bb44de6758dd14a93722098d35fc028a5e245e09d9874da5"} Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.153212 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e2caef7fbe2e8f4bb44de6758dd14a93722098d35fc028a5e245e09d9874da5" Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.167637 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1126d25d-8501-4fd5-9ac6-cde91e58ac59","Type":"ContainerStarted","Data":"239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703"} Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.167843 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1126d25d-8501-4fd5-9ac6-cde91e58ac59" containerName="cinder-api-log" containerID="cri-o://0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e" gracePeriod=30 Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.167955 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.168410 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1126d25d-8501-4fd5-9ac6-cde91e58ac59" containerName="cinder-api" containerID="cri-o://239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703" gracePeriod=30 Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.175115 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8536785-c783-4780-9135-4a460f274c68","Type":"ContainerStarted","Data":"e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a"} Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.194109 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.194081828 podStartE2EDuration="5.194081828s" podCreationTimestamp="2025-10-08 09:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:29:43.191422606 +0000 UTC m=+1078.439067845" watchObservedRunningTime="2025-10-08 09:29:43.194081828 +0000 UTC m=+1078.441727067" Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.523093 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-plnkl" Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.661243 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78khw\" (UniqueName: \"kubernetes.io/projected/929c71a8-07ed-475b-b150-e5a1d92e22c5-kube-api-access-78khw\") pod \"929c71a8-07ed-475b-b150-e5a1d92e22c5\" (UID: \"929c71a8-07ed-475b-b150-e5a1d92e22c5\") " Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.668711 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/929c71a8-07ed-475b-b150-e5a1d92e22c5-kube-api-access-78khw" (OuterVolumeSpecName: "kube-api-access-78khw") pod "929c71a8-07ed-475b-b150-e5a1d92e22c5" (UID: "929c71a8-07ed-475b-b150-e5a1d92e22c5"). InnerVolumeSpecName "kube-api-access-78khw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:43 crc kubenswrapper[4744]: I1008 09:29:43.764591 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78khw\" (UniqueName: \"kubernetes.io/projected/929c71a8-07ed-475b-b150-e5a1d92e22c5-kube-api-access-78khw\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.146772 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.191718 4744 generic.go:334] "Generic (PLEG): container finished" podID="1126d25d-8501-4fd5-9ac6-cde91e58ac59" containerID="239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703" exitCode=0 Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.191765 4744 generic.go:334] "Generic (PLEG): container finished" podID="1126d25d-8501-4fd5-9ac6-cde91e58ac59" containerID="0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e" exitCode=143 Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.191820 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1126d25d-8501-4fd5-9ac6-cde91e58ac59","Type":"ContainerDied","Data":"239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703"} Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.191861 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1126d25d-8501-4fd5-9ac6-cde91e58ac59","Type":"ContainerDied","Data":"0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e"} Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.191876 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1126d25d-8501-4fd5-9ac6-cde91e58ac59","Type":"ContainerDied","Data":"8ed9ad67eaa32d8ae1bca1dd2c5f2946008c10628211b7ceb47aa68f2fa15d7a"} Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.191899 4744 scope.go:117] "RemoveContainer" containerID="239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.192107 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.199700 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8536785-c783-4780-9135-4a460f274c68","Type":"ContainerStarted","Data":"f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540"} Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.200008 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.205169 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8a257281-2eda-4f01-920f-02757ef0fb37","Type":"ContainerStarted","Data":"ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659"} Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.215141 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-plnkl" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.217512 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-plnkl" event={"ID":"929c71a8-07ed-475b-b150-e5a1d92e22c5","Type":"ContainerDied","Data":"1098ce5d3939831583b980dfaa72dd6f675c376418ca6dc821fa77d31e2b1417"} Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.217595 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1098ce5d3939831583b980dfaa72dd6f675c376418ca6dc821fa77d31e2b1417" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.240392 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.7098756760000002 podStartE2EDuration="7.240348157s" podCreationTimestamp="2025-10-08 09:29:37 +0000 UTC" firstStartedPulling="2025-10-08 09:29:39.07634002 +0000 UTC m=+1074.323985259" lastFinishedPulling="2025-10-08 09:29:43.606812501 +0000 UTC m=+1078.854457740" observedRunningTime="2025-10-08 09:29:44.229029494 +0000 UTC m=+1079.476674763" watchObservedRunningTime="2025-10-08 09:29:44.240348157 +0000 UTC m=+1079.487993396" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.269830 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.8521330670000005 podStartE2EDuration="6.269807645s" podCreationTimestamp="2025-10-08 09:29:38 +0000 UTC" firstStartedPulling="2025-10-08 09:29:39.66884689 +0000 UTC m=+1074.916492119" lastFinishedPulling="2025-10-08 09:29:41.086521458 +0000 UTC m=+1076.334166697" observedRunningTime="2025-10-08 09:29:44.266036274 +0000 UTC m=+1079.513681513" watchObservedRunningTime="2025-10-08 09:29:44.269807645 +0000 UTC m=+1079.517452884" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.277519 4744 scope.go:117] "RemoveContainer" containerID="0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.284205 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1126d25d-8501-4fd5-9ac6-cde91e58ac59-logs\") pod \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.284278 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhvr7\" (UniqueName: \"kubernetes.io/projected/1126d25d-8501-4fd5-9ac6-cde91e58ac59-kube-api-access-nhvr7\") pod \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.284445 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-combined-ca-bundle\") pod \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.284496 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-scripts\") pod \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.284527 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1126d25d-8501-4fd5-9ac6-cde91e58ac59-etc-machine-id\") pod \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.284572 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-config-data-custom\") pod \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.284691 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-config-data\") pod \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\" (UID: \"1126d25d-8501-4fd5-9ac6-cde91e58ac59\") " Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.294607 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1126d25d-8501-4fd5-9ac6-cde91e58ac59-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1126d25d-8501-4fd5-9ac6-cde91e58ac59" (UID: "1126d25d-8501-4fd5-9ac6-cde91e58ac59"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.295413 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1126d25d-8501-4fd5-9ac6-cde91e58ac59-logs" (OuterVolumeSpecName: "logs") pod "1126d25d-8501-4fd5-9ac6-cde91e58ac59" (UID: "1126d25d-8501-4fd5-9ac6-cde91e58ac59"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.309808 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1126d25d-8501-4fd5-9ac6-cde91e58ac59-kube-api-access-nhvr7" (OuterVolumeSpecName: "kube-api-access-nhvr7") pod "1126d25d-8501-4fd5-9ac6-cde91e58ac59" (UID: "1126d25d-8501-4fd5-9ac6-cde91e58ac59"). InnerVolumeSpecName "kube-api-access-nhvr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.328973 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-scripts" (OuterVolumeSpecName: "scripts") pod "1126d25d-8501-4fd5-9ac6-cde91e58ac59" (UID: "1126d25d-8501-4fd5-9ac6-cde91e58ac59"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.335416 4744 scope.go:117] "RemoveContainer" containerID="239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703" Oct 08 09:29:44 crc kubenswrapper[4744]: E1008 09:29:44.341525 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703\": container with ID starting with 239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703 not found: ID does not exist" containerID="239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.341600 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703"} err="failed to get container status \"239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703\": rpc error: code = NotFound desc = could not find container \"239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703\": container with ID starting with 239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703 not found: ID does not exist" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.341641 4744 scope.go:117] "RemoveContainer" containerID="0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e" Oct 08 09:29:44 crc kubenswrapper[4744]: E1008 09:29:44.346535 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e\": container with ID starting with 0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e not found: ID does not exist" containerID="0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.346582 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e"} err="failed to get container status \"0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e\": rpc error: code = NotFound desc = could not find container \"0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e\": container with ID starting with 0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e not found: ID does not exist" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.346614 4744 scope.go:117] "RemoveContainer" containerID="239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.354465 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703"} err="failed to get container status \"239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703\": rpc error: code = NotFound desc = could not find container \"239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703\": container with ID starting with 239cd05f059cdf2dffb8cd694e9a63c6f469b914031b51b76bf35b9da8803703 not found: ID does not exist" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.354523 4744 scope.go:117] "RemoveContainer" containerID="0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.359099 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e"} err="failed to get container status \"0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e\": rpc error: code = NotFound desc = could not find container \"0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e\": container with ID starting with 0353f244d43c8fe3d843138e0db072edf3f1faa2c85fec9145a8c0bc8ff8f42e not found: ID does not exist" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.364132 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1126d25d-8501-4fd5-9ac6-cde91e58ac59" (UID: "1126d25d-8501-4fd5-9ac6-cde91e58ac59"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.396524 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.396559 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1126d25d-8501-4fd5-9ac6-cde91e58ac59-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.396569 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.396580 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1126d25d-8501-4fd5-9ac6-cde91e58ac59-logs\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.396588 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhvr7\" (UniqueName: \"kubernetes.io/projected/1126d25d-8501-4fd5-9ac6-cde91e58ac59-kube-api-access-nhvr7\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.399606 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1126d25d-8501-4fd5-9ac6-cde91e58ac59" (UID: "1126d25d-8501-4fd5-9ac6-cde91e58ac59"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.434522 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-config-data" (OuterVolumeSpecName: "config-data") pod "1126d25d-8501-4fd5-9ac6-cde91e58ac59" (UID: "1126d25d-8501-4fd5-9ac6-cde91e58ac59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.498754 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.498802 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1126d25d-8501-4fd5-9ac6-cde91e58ac59-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.550026 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.566464 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.602006 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 09:29:44 crc kubenswrapper[4744]: E1008 09:29:44.602516 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="929c71a8-07ed-475b-b150-e5a1d92e22c5" containerName="mariadb-database-create" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.602539 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="929c71a8-07ed-475b-b150-e5a1d92e22c5" containerName="mariadb-database-create" Oct 08 09:29:44 crc kubenswrapper[4744]: E1008 09:29:44.602563 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1126d25d-8501-4fd5-9ac6-cde91e58ac59" containerName="cinder-api-log" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.602570 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1126d25d-8501-4fd5-9ac6-cde91e58ac59" containerName="cinder-api-log" Oct 08 09:29:44 crc kubenswrapper[4744]: E1008 09:29:44.602589 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56496cc2-c5e8-426d-8f7c-fbd84f7b266c" containerName="mariadb-database-create" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.602595 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="56496cc2-c5e8-426d-8f7c-fbd84f7b266c" containerName="mariadb-database-create" Oct 08 09:29:44 crc kubenswrapper[4744]: E1008 09:29:44.602608 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1126d25d-8501-4fd5-9ac6-cde91e58ac59" containerName="cinder-api" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.602615 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1126d25d-8501-4fd5-9ac6-cde91e58ac59" containerName="cinder-api" Oct 08 09:29:44 crc kubenswrapper[4744]: E1008 09:29:44.602629 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daa62f04-afb7-4d1e-9218-cbc2cc22e807" containerName="mariadb-database-create" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.602635 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="daa62f04-afb7-4d1e-9218-cbc2cc22e807" containerName="mariadb-database-create" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.602840 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1126d25d-8501-4fd5-9ac6-cde91e58ac59" containerName="cinder-api-log" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.602858 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1126d25d-8501-4fd5-9ac6-cde91e58ac59" containerName="cinder-api" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.602881 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="daa62f04-afb7-4d1e-9218-cbc2cc22e807" containerName="mariadb-database-create" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.602895 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="929c71a8-07ed-475b-b150-e5a1d92e22c5" containerName="mariadb-database-create" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.602915 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="56496cc2-c5e8-426d-8f7c-fbd84f7b266c" containerName="mariadb-database-create" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.607287 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.621467 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.624922 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.626221 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.638969 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.707721 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-scripts\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.707787 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9efda746-3b68-4707-bbe8-3f4e74db5a60-logs\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.707816 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-config-data\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.707839 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9efda746-3b68-4707-bbe8-3f4e74db5a60-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.708016 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km24w\" (UniqueName: \"kubernetes.io/projected/9efda746-3b68-4707-bbe8-3f4e74db5a60-kube-api-access-km24w\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.708150 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.708256 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-config-data-custom\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.710130 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.710214 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.811638 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.812143 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-scripts\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.812176 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9efda746-3b68-4707-bbe8-3f4e74db5a60-logs\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.812200 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-config-data\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.812225 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9efda746-3b68-4707-bbe8-3f4e74db5a60-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.812259 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km24w\" (UniqueName: \"kubernetes.io/projected/9efda746-3b68-4707-bbe8-3f4e74db5a60-kube-api-access-km24w\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.812285 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.812319 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-config-data-custom\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.812439 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.812612 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9efda746-3b68-4707-bbe8-3f4e74db5a60-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.813916 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9efda746-3b68-4707-bbe8-3f4e74db5a60-logs\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.821506 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.821589 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.822245 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-config-data-custom\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.825607 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-scripts\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.827887 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-config-data\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.838070 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9efda746-3b68-4707-bbe8-3f4e74db5a60-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:44 crc kubenswrapper[4744]: I1008 09:29:44.879081 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km24w\" (UniqueName: \"kubernetes.io/projected/9efda746-3b68-4707-bbe8-3f4e74db5a60-kube-api-access-km24w\") pod \"cinder-api-0\" (UID: \"9efda746-3b68-4707-bbe8-3f4e74db5a60\") " pod="openstack/cinder-api-0" Oct 08 09:29:45 crc kubenswrapper[4744]: I1008 09:29:45.026405 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 09:29:45 crc kubenswrapper[4744]: I1008 09:29:45.487454 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1126d25d-8501-4fd5-9ac6-cde91e58ac59" path="/var/lib/kubelet/pods/1126d25d-8501-4fd5-9ac6-cde91e58ac59/volumes" Oct 08 09:29:45 crc kubenswrapper[4744]: I1008 09:29:45.720665 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 09:29:46 crc kubenswrapper[4744]: I1008 09:29:46.255313 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9efda746-3b68-4707-bbe8-3f4e74db5a60","Type":"ContainerStarted","Data":"6f5dc0bf3a9be408660098f2b1d7d9190d510f26f4b8a8fcd77e2b31c69c3479"} Oct 08 09:29:46 crc kubenswrapper[4744]: I1008 09:29:46.871101 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:46 crc kubenswrapper[4744]: I1008 09:29:46.876265 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="ceilometer-central-agent" containerID="cri-o://f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5" gracePeriod=30 Oct 08 09:29:46 crc kubenswrapper[4744]: I1008 09:29:46.877052 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="proxy-httpd" containerID="cri-o://f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540" gracePeriod=30 Oct 08 09:29:46 crc kubenswrapper[4744]: I1008 09:29:46.877119 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="sg-core" containerID="cri-o://e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a" gracePeriod=30 Oct 08 09:29:46 crc kubenswrapper[4744]: I1008 09:29:46.877172 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="ceilometer-notification-agent" containerID="cri-o://6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d" gracePeriod=30 Oct 08 09:29:47 crc kubenswrapper[4744]: I1008 09:29:47.276008 4744 generic.go:334] "Generic (PLEG): container finished" podID="a8536785-c783-4780-9135-4a460f274c68" containerID="f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540" exitCode=0 Oct 08 09:29:47 crc kubenswrapper[4744]: I1008 09:29:47.276632 4744 generic.go:334] "Generic (PLEG): container finished" podID="a8536785-c783-4780-9135-4a460f274c68" containerID="e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a" exitCode=2 Oct 08 09:29:47 crc kubenswrapper[4744]: I1008 09:29:47.276420 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8536785-c783-4780-9135-4a460f274c68","Type":"ContainerDied","Data":"f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540"} Oct 08 09:29:47 crc kubenswrapper[4744]: I1008 09:29:47.277722 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8536785-c783-4780-9135-4a460f274c68","Type":"ContainerDied","Data":"e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a"} Oct 08 09:29:47 crc kubenswrapper[4744]: I1008 09:29:47.283482 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9efda746-3b68-4707-bbe8-3f4e74db5a60","Type":"ContainerStarted","Data":"ebfa3ab65af1537f046e19a6c513a3163857ff2fe07f9492afa70f200fec65c8"} Oct 08 09:29:47 crc kubenswrapper[4744]: I1008 09:29:47.983656 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.102319 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-sg-core-conf-yaml\") pod \"a8536785-c783-4780-9135-4a460f274c68\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.102424 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8536785-c783-4780-9135-4a460f274c68-log-httpd\") pod \"a8536785-c783-4780-9135-4a460f274c68\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.102466 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-config-data\") pod \"a8536785-c783-4780-9135-4a460f274c68\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.102525 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqz57\" (UniqueName: \"kubernetes.io/projected/a8536785-c783-4780-9135-4a460f274c68-kube-api-access-fqz57\") pod \"a8536785-c783-4780-9135-4a460f274c68\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.102568 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-combined-ca-bundle\") pod \"a8536785-c783-4780-9135-4a460f274c68\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.102643 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-scripts\") pod \"a8536785-c783-4780-9135-4a460f274c68\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.102800 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8536785-c783-4780-9135-4a460f274c68-run-httpd\") pod \"a8536785-c783-4780-9135-4a460f274c68\" (UID: \"a8536785-c783-4780-9135-4a460f274c68\") " Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.103955 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8536785-c783-4780-9135-4a460f274c68-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a8536785-c783-4780-9135-4a460f274c68" (UID: "a8536785-c783-4780-9135-4a460f274c68"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.103986 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8536785-c783-4780-9135-4a460f274c68-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a8536785-c783-4780-9135-4a460f274c68" (UID: "a8536785-c783-4780-9135-4a460f274c68"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.113864 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-scripts" (OuterVolumeSpecName: "scripts") pod "a8536785-c783-4780-9135-4a460f274c68" (UID: "a8536785-c783-4780-9135-4a460f274c68"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.134346 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8536785-c783-4780-9135-4a460f274c68-kube-api-access-fqz57" (OuterVolumeSpecName: "kube-api-access-fqz57") pod "a8536785-c783-4780-9135-4a460f274c68" (UID: "a8536785-c783-4780-9135-4a460f274c68"). InnerVolumeSpecName "kube-api-access-fqz57". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.149336 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a8536785-c783-4780-9135-4a460f274c68" (UID: "a8536785-c783-4780-9135-4a460f274c68"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.206096 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8536785-c783-4780-9135-4a460f274c68-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.206136 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.206149 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8536785-c783-4780-9135-4a460f274c68-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.206160 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqz57\" (UniqueName: \"kubernetes.io/projected/a8536785-c783-4780-9135-4a460f274c68-kube-api-access-fqz57\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.206168 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.220361 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8536785-c783-4780-9135-4a460f274c68" (UID: "a8536785-c783-4780-9135-4a460f274c68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.275928 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-config-data" (OuterVolumeSpecName: "config-data") pod "a8536785-c783-4780-9135-4a460f274c68" (UID: "a8536785-c783-4780-9135-4a460f274c68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.297797 4744 generic.go:334] "Generic (PLEG): container finished" podID="a8536785-c783-4780-9135-4a460f274c68" containerID="6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d" exitCode=0 Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.297851 4744 generic.go:334] "Generic (PLEG): container finished" podID="a8536785-c783-4780-9135-4a460f274c68" containerID="f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5" exitCode=0 Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.297889 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.298289 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8536785-c783-4780-9135-4a460f274c68","Type":"ContainerDied","Data":"6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d"} Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.298474 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8536785-c783-4780-9135-4a460f274c68","Type":"ContainerDied","Data":"f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5"} Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.298566 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8536785-c783-4780-9135-4a460f274c68","Type":"ContainerDied","Data":"09e3cdc99ecfcdaab4e8ae30f37b0982c0d1a73274e978e0b59df8f036431f08"} Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.298513 4744 scope.go:117] "RemoveContainer" containerID="f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.301134 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9efda746-3b68-4707-bbe8-3f4e74db5a60","Type":"ContainerStarted","Data":"4f74d59d19efc867109b5117a2615ee16178ca3e5065b5b9714fe4a791b0e59e"} Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.302642 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.308685 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.308840 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8536785-c783-4780-9135-4a460f274c68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.325868 4744 scope.go:117] "RemoveContainer" containerID="e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.336493 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.336471247 podStartE2EDuration="4.336471247s" podCreationTimestamp="2025-10-08 09:29:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:29:48.332799958 +0000 UTC m=+1083.580445197" watchObservedRunningTime="2025-10-08 09:29:48.336471247 +0000 UTC m=+1083.584116486" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.379253 4744 scope.go:117] "RemoveContainer" containerID="6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.379443 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.401685 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.408326 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:48 crc kubenswrapper[4744]: E1008 09:29:48.408828 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="proxy-httpd" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.408848 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="proxy-httpd" Oct 08 09:29:48 crc kubenswrapper[4744]: E1008 09:29:48.408876 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="sg-core" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.408883 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="sg-core" Oct 08 09:29:48 crc kubenswrapper[4744]: E1008 09:29:48.408908 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="ceilometer-notification-agent" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.408916 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="ceilometer-notification-agent" Oct 08 09:29:48 crc kubenswrapper[4744]: E1008 09:29:48.408928 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="ceilometer-central-agent" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.408934 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="ceilometer-central-agent" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.409121 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="ceilometer-notification-agent" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.409138 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="sg-core" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.409157 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="proxy-httpd" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.409169 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8536785-c783-4780-9135-4a460f274c68" containerName="ceilometer-central-agent" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.410825 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.416960 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.417198 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.422001 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.428188 4744 scope.go:117] "RemoveContainer" containerID="f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.464594 4744 scope.go:117] "RemoveContainer" containerID="f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540" Oct 08 09:29:48 crc kubenswrapper[4744]: E1008 09:29:48.465433 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540\": container with ID starting with f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540 not found: ID does not exist" containerID="f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.465480 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540"} err="failed to get container status \"f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540\": rpc error: code = NotFound desc = could not find container \"f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540\": container with ID starting with f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540 not found: ID does not exist" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.465507 4744 scope.go:117] "RemoveContainer" containerID="e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a" Oct 08 09:29:48 crc kubenswrapper[4744]: E1008 09:29:48.467569 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a\": container with ID starting with e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a not found: ID does not exist" containerID="e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.467597 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a"} err="failed to get container status \"e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a\": rpc error: code = NotFound desc = could not find container \"e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a\": container with ID starting with e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a not found: ID does not exist" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.467613 4744 scope.go:117] "RemoveContainer" containerID="6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d" Oct 08 09:29:48 crc kubenswrapper[4744]: E1008 09:29:48.469141 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d\": container with ID starting with 6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d not found: ID does not exist" containerID="6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.469225 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d"} err="failed to get container status \"6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d\": rpc error: code = NotFound desc = could not find container \"6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d\": container with ID starting with 6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d not found: ID does not exist" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.469279 4744 scope.go:117] "RemoveContainer" containerID="f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5" Oct 08 09:29:48 crc kubenswrapper[4744]: E1008 09:29:48.469994 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5\": container with ID starting with f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5 not found: ID does not exist" containerID="f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.470068 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5"} err="failed to get container status \"f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5\": rpc error: code = NotFound desc = could not find container \"f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5\": container with ID starting with f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5 not found: ID does not exist" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.470123 4744 scope.go:117] "RemoveContainer" containerID="f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.470516 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540"} err="failed to get container status \"f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540\": rpc error: code = NotFound desc = could not find container \"f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540\": container with ID starting with f51579c6ab40abfebf7ae1588412814f992189802667ed93ed119162a48bb540 not found: ID does not exist" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.470540 4744 scope.go:117] "RemoveContainer" containerID="e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.471438 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a"} err="failed to get container status \"e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a\": rpc error: code = NotFound desc = could not find container \"e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a\": container with ID starting with e9ae1f8efe23aad8bba1407b5a78639726de890cce093d39313bc69b10d8317a not found: ID does not exist" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.471488 4744 scope.go:117] "RemoveContainer" containerID="6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.473177 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d"} err="failed to get container status \"6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d\": rpc error: code = NotFound desc = could not find container \"6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d\": container with ID starting with 6625a838b9c78101c873ec5a39c4f508da3441f35c275579258c9a800992710d not found: ID does not exist" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.473240 4744 scope.go:117] "RemoveContainer" containerID="f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.473667 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5"} err="failed to get container status \"f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5\": rpc error: code = NotFound desc = could not find container \"f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5\": container with ID starting with f17aee8ea844e8a707176b10c41d2d2b4177cc95b9f26acf365406fe91adfab5 not found: ID does not exist" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.505534 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.513006 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-run-httpd\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.513098 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.513123 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-log-httpd\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.513981 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-config-data\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.514025 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5gpk\" (UniqueName: \"kubernetes.io/projected/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-kube-api-access-h5gpk\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.514044 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.514120 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-scripts\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.615604 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-scripts\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.615665 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-run-httpd\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.615729 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.615762 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-log-httpd\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.615843 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-config-data\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.615866 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5gpk\" (UniqueName: \"kubernetes.io/projected/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-kube-api-access-h5gpk\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.615893 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.618020 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-log-httpd\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.618081 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-run-httpd\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.622258 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-scripts\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.622924 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.625289 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-config-data\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.632335 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.668116 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5gpk\" (UniqueName: \"kubernetes.io/projected/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-kube-api-access-h5gpk\") pod \"ceilometer-0\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.742352 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.802448 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.817602 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.966760 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5bdc4b9-8znmp"] Oct 08 09:29:48 crc kubenswrapper[4744]: I1008 09:29:48.967475 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" podUID="8306420e-9e3d-48a6-87b9-af6c61f0d0ba" containerName="dnsmasq-dns" containerID="cri-o://0df1f11b6f6ca22fe95cb08ac215a75ef6b758319ee0d0242c465c55916d7ae6" gracePeriod=10 Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.041424 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-bc72-account-create-gfwls"] Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.042655 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bc72-account-create-gfwls" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.049499 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.066972 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-bc72-account-create-gfwls"] Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.144868 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7c5j\" (UniqueName: \"kubernetes.io/projected/2edcd16b-76ad-4a92-b75b-033b9edd0024-kube-api-access-r7c5j\") pod \"nova-api-bc72-account-create-gfwls\" (UID: \"2edcd16b-76ad-4a92-b75b-033b9edd0024\") " pod="openstack/nova-api-bc72-account-create-gfwls" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.245474 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-9b88-account-create-v6x5w"] Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.247207 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b88-account-create-v6x5w" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.251738 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.253825 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7c5j\" (UniqueName: \"kubernetes.io/projected/2edcd16b-76ad-4a92-b75b-033b9edd0024-kube-api-access-r7c5j\") pod \"nova-api-bc72-account-create-gfwls\" (UID: \"2edcd16b-76ad-4a92-b75b-033b9edd0024\") " pod="openstack/nova-api-bc72-account-create-gfwls" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.263485 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9b88-account-create-v6x5w"] Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.291002 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7c5j\" (UniqueName: \"kubernetes.io/projected/2edcd16b-76ad-4a92-b75b-033b9edd0024-kube-api-access-r7c5j\") pod \"nova-api-bc72-account-create-gfwls\" (UID: \"2edcd16b-76ad-4a92-b75b-033b9edd0024\") " pod="openstack/nova-api-bc72-account-create-gfwls" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.324524 4744 generic.go:334] "Generic (PLEG): container finished" podID="8306420e-9e3d-48a6-87b9-af6c61f0d0ba" containerID="0df1f11b6f6ca22fe95cb08ac215a75ef6b758319ee0d0242c465c55916d7ae6" exitCode=0 Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.324658 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" event={"ID":"8306420e-9e3d-48a6-87b9-af6c61f0d0ba","Type":"ContainerDied","Data":"0df1f11b6f6ca22fe95cb08ac215a75ef6b758319ee0d0242c465c55916d7ae6"} Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.356436 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktmhs\" (UniqueName: \"kubernetes.io/projected/7edf84ea-4d5a-4764-b6a7-ea03f6e312e4-kube-api-access-ktmhs\") pod \"nova-cell0-9b88-account-create-v6x5w\" (UID: \"7edf84ea-4d5a-4764-b6a7-ea03f6e312e4\") " pod="openstack/nova-cell0-9b88-account-create-v6x5w" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.378248 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bc72-account-create-gfwls" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.409407 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.424502 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.459419 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktmhs\" (UniqueName: \"kubernetes.io/projected/7edf84ea-4d5a-4764-b6a7-ea03f6e312e4-kube-api-access-ktmhs\") pod \"nova-cell0-9b88-account-create-v6x5w\" (UID: \"7edf84ea-4d5a-4764-b6a7-ea03f6e312e4\") " pod="openstack/nova-cell0-9b88-account-create-v6x5w" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.476754 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8536785-c783-4780-9135-4a460f274c68" path="/var/lib/kubelet/pods/a8536785-c783-4780-9135-4a460f274c68/volumes" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.483791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktmhs\" (UniqueName: \"kubernetes.io/projected/7edf84ea-4d5a-4764-b6a7-ea03f6e312e4-kube-api-access-ktmhs\") pod \"nova-cell0-9b88-account-create-v6x5w\" (UID: \"7edf84ea-4d5a-4764-b6a7-ea03f6e312e4\") " pod="openstack/nova-cell0-9b88-account-create-v6x5w" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.568900 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b88-account-create-v6x5w" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.618164 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.665481 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-ovsdbserver-sb\") pod \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.665572 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-dns-svc\") pod \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.665631 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-ovsdbserver-nb\") pod \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.665690 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-config\") pod \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.665759 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fztsc\" (UniqueName: \"kubernetes.io/projected/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-kube-api-access-fztsc\") pod \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\" (UID: \"8306420e-9e3d-48a6-87b9-af6c61f0d0ba\") " Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.709556 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-kube-api-access-fztsc" (OuterVolumeSpecName: "kube-api-access-fztsc") pod "8306420e-9e3d-48a6-87b9-af6c61f0d0ba" (UID: "8306420e-9e3d-48a6-87b9-af6c61f0d0ba"). InnerVolumeSpecName "kube-api-access-fztsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.709718 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.709774 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.770912 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fztsc\" (UniqueName: \"kubernetes.io/projected/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-kube-api-access-fztsc\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.795461 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8306420e-9e3d-48a6-87b9-af6c61f0d0ba" (UID: "8306420e-9e3d-48a6-87b9-af6c61f0d0ba"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.829050 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8306420e-9e3d-48a6-87b9-af6c61f0d0ba" (UID: "8306420e-9e3d-48a6-87b9-af6c61f0d0ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.837286 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8306420e-9e3d-48a6-87b9-af6c61f0d0ba" (UID: "8306420e-9e3d-48a6-87b9-af6c61f0d0ba"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.876004 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.876031 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.876041 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.882583 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-config" (OuterVolumeSpecName: "config") pod "8306420e-9e3d-48a6-87b9-af6c61f0d0ba" (UID: "8306420e-9e3d-48a6-87b9-af6c61f0d0ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:29:49 crc kubenswrapper[4744]: I1008 09:29:49.979638 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8306420e-9e3d-48a6-87b9-af6c61f0d0ba-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.068864 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-bc72-account-create-gfwls"] Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.334978 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9b88-account-create-v6x5w"] Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.345211 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d","Type":"ContainerStarted","Data":"10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245"} Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.345270 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d","Type":"ContainerStarted","Data":"c1dfe6c2358b18d1d80ad4a53c669a6f1526d3cc838e4dced863870b58a27000"} Oct 08 09:29:50 crc kubenswrapper[4744]: W1008 09:29:50.347609 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7edf84ea_4d5a_4764_b6a7_ea03f6e312e4.slice/crio-dc2bbe45ef6bf258035593e0805d687de532def819071ba206994a6075f76157 WatchSource:0}: Error finding container dc2bbe45ef6bf258035593e0805d687de532def819071ba206994a6075f76157: Status 404 returned error can't find the container with id dc2bbe45ef6bf258035593e0805d687de532def819071ba206994a6075f76157 Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.347945 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" event={"ID":"8306420e-9e3d-48a6-87b9-af6c61f0d0ba","Type":"ContainerDied","Data":"b3531898dd0b6c3170e6437b94f43f0f4d0d1280bf6f9ced9fa18d2125c8f25f"} Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.347985 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5bdc4b9-8znmp" Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.348001 4744 scope.go:117] "RemoveContainer" containerID="0df1f11b6f6ca22fe95cb08ac215a75ef6b758319ee0d0242c465c55916d7ae6" Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.358181 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8a257281-2eda-4f01-920f-02757ef0fb37" containerName="cinder-scheduler" containerID="cri-o://3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880" gracePeriod=30 Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.358311 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bc72-account-create-gfwls" event={"ID":"2edcd16b-76ad-4a92-b75b-033b9edd0024","Type":"ContainerStarted","Data":"d855e82220b5b6051eab40e717ea5cb287ad149342b83df5f6a118f2957be4f5"} Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.358731 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8a257281-2eda-4f01-920f-02757ef0fb37" containerName="probe" containerID="cri-o://ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659" gracePeriod=30 Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.419463 4744 scope.go:117] "RemoveContainer" containerID="d0f22e1597065354a52bdc3836dabe667b45b928fc1bc324cf462d7cd4f0f325" Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.426429 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5bdc4b9-8znmp"] Oct 08 09:29:50 crc kubenswrapper[4744]: I1008 09:29:50.445095 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5bdc4b9-8znmp"] Oct 08 09:29:51 crc kubenswrapper[4744]: I1008 09:29:51.371967 4744 generic.go:334] "Generic (PLEG): container finished" podID="2edcd16b-76ad-4a92-b75b-033b9edd0024" containerID="01ab5fc768cb62b708dd9ca19f8ff741686c711b9bdd4095e355b80c2000bde9" exitCode=0 Oct 08 09:29:51 crc kubenswrapper[4744]: I1008 09:29:51.372078 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bc72-account-create-gfwls" event={"ID":"2edcd16b-76ad-4a92-b75b-033b9edd0024","Type":"ContainerDied","Data":"01ab5fc768cb62b708dd9ca19f8ff741686c711b9bdd4095e355b80c2000bde9"} Oct 08 09:29:51 crc kubenswrapper[4744]: I1008 09:29:51.375637 4744 generic.go:334] "Generic (PLEG): container finished" podID="7edf84ea-4d5a-4764-b6a7-ea03f6e312e4" containerID="5aa01bfa090d80a1d857280f83be8e938f84d2155b94a373a4aa6fea1e452704" exitCode=0 Oct 08 09:29:51 crc kubenswrapper[4744]: I1008 09:29:51.375699 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b88-account-create-v6x5w" event={"ID":"7edf84ea-4d5a-4764-b6a7-ea03f6e312e4","Type":"ContainerDied","Data":"5aa01bfa090d80a1d857280f83be8e938f84d2155b94a373a4aa6fea1e452704"} Oct 08 09:29:51 crc kubenswrapper[4744]: I1008 09:29:51.375727 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b88-account-create-v6x5w" event={"ID":"7edf84ea-4d5a-4764-b6a7-ea03f6e312e4","Type":"ContainerStarted","Data":"dc2bbe45ef6bf258035593e0805d687de532def819071ba206994a6075f76157"} Oct 08 09:29:51 crc kubenswrapper[4744]: I1008 09:29:51.380656 4744 generic.go:334] "Generic (PLEG): container finished" podID="8a257281-2eda-4f01-920f-02757ef0fb37" containerID="ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659" exitCode=0 Oct 08 09:29:51 crc kubenswrapper[4744]: I1008 09:29:51.380718 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8a257281-2eda-4f01-920f-02757ef0fb37","Type":"ContainerDied","Data":"ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659"} Oct 08 09:29:51 crc kubenswrapper[4744]: I1008 09:29:51.383926 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d","Type":"ContainerStarted","Data":"514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6"} Oct 08 09:29:51 crc kubenswrapper[4744]: I1008 09:29:51.524654 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8306420e-9e3d-48a6-87b9-af6c61f0d0ba" path="/var/lib/kubelet/pods/8306420e-9e3d-48a6-87b9-af6c61f0d0ba/volumes" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.285753 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.375383 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.397988 4744 generic.go:334] "Generic (PLEG): container finished" podID="8a257281-2eda-4f01-920f-02757ef0fb37" containerID="3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880" exitCode=0 Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.398071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8a257281-2eda-4f01-920f-02757ef0fb37","Type":"ContainerDied","Data":"3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880"} Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.398111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8a257281-2eda-4f01-920f-02757ef0fb37","Type":"ContainerDied","Data":"8159019b86efb3037ce6bc5db6141800383b41c1982f22c60b4fa072bfb01d13"} Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.398134 4744 scope.go:117] "RemoveContainer" containerID="ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.398277 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.412259 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d","Type":"ContainerStarted","Data":"1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95"} Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.443027 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-combined-ca-bundle\") pod \"8a257281-2eda-4f01-920f-02757ef0fb37\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.443279 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-scripts\") pod \"8a257281-2eda-4f01-920f-02757ef0fb37\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.443355 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhtsp\" (UniqueName: \"kubernetes.io/projected/8a257281-2eda-4f01-920f-02757ef0fb37-kube-api-access-lhtsp\") pod \"8a257281-2eda-4f01-920f-02757ef0fb37\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.443400 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a257281-2eda-4f01-920f-02757ef0fb37-etc-machine-id\") pod \"8a257281-2eda-4f01-920f-02757ef0fb37\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.443454 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-config-data-custom\") pod \"8a257281-2eda-4f01-920f-02757ef0fb37\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.443495 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-config-data\") pod \"8a257281-2eda-4f01-920f-02757ef0fb37\" (UID: \"8a257281-2eda-4f01-920f-02757ef0fb37\") " Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.444097 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a257281-2eda-4f01-920f-02757ef0fb37-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8a257281-2eda-4f01-920f-02757ef0fb37" (UID: "8a257281-2eda-4f01-920f-02757ef0fb37"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.454959 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-scripts" (OuterVolumeSpecName: "scripts") pod "8a257281-2eda-4f01-920f-02757ef0fb37" (UID: "8a257281-2eda-4f01-920f-02757ef0fb37"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.477490 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8a257281-2eda-4f01-920f-02757ef0fb37" (UID: "8a257281-2eda-4f01-920f-02757ef0fb37"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.477694 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a257281-2eda-4f01-920f-02757ef0fb37-kube-api-access-lhtsp" (OuterVolumeSpecName: "kube-api-access-lhtsp") pod "8a257281-2eda-4f01-920f-02757ef0fb37" (UID: "8a257281-2eda-4f01-920f-02757ef0fb37"). InnerVolumeSpecName "kube-api-access-lhtsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.488592 4744 scope.go:117] "RemoveContainer" containerID="3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.546787 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.546819 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhtsp\" (UniqueName: \"kubernetes.io/projected/8a257281-2eda-4f01-920f-02757ef0fb37-kube-api-access-lhtsp\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.546832 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a257281-2eda-4f01-920f-02757ef0fb37-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.546841 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.568689 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a257281-2eda-4f01-920f-02757ef0fb37" (UID: "8a257281-2eda-4f01-920f-02757ef0fb37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.574157 4744 scope.go:117] "RemoveContainer" containerID="ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659" Oct 08 09:29:52 crc kubenswrapper[4744]: E1008 09:29:52.574904 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659\": container with ID starting with ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659 not found: ID does not exist" containerID="ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.574967 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659"} err="failed to get container status \"ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659\": rpc error: code = NotFound desc = could not find container \"ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659\": container with ID starting with ee11069b5c3f2e2b441d9d8d8358651f283e2f57c1c5d697853f550ca7c77659 not found: ID does not exist" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.575002 4744 scope.go:117] "RemoveContainer" containerID="3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880" Oct 08 09:29:52 crc kubenswrapper[4744]: E1008 09:29:52.577852 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880\": container with ID starting with 3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880 not found: ID does not exist" containerID="3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.577914 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880"} err="failed to get container status \"3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880\": rpc error: code = NotFound desc = could not find container \"3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880\": container with ID starting with 3f48abfbf720447a1fe4013b9d76e07f30717dc215cfe2b1d720c6cb2e849880 not found: ID does not exist" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.634316 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-config-data" (OuterVolumeSpecName: "config-data") pod "8a257281-2eda-4f01-920f-02757ef0fb37" (UID: "8a257281-2eda-4f01-920f-02757ef0fb37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.649368 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.649406 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a257281-2eda-4f01-920f-02757ef0fb37-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.790472 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.798267 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.853827 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 09:29:52 crc kubenswrapper[4744]: E1008 09:29:52.854448 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8306420e-9e3d-48a6-87b9-af6c61f0d0ba" containerName="dnsmasq-dns" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.854468 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8306420e-9e3d-48a6-87b9-af6c61f0d0ba" containerName="dnsmasq-dns" Oct 08 09:29:52 crc kubenswrapper[4744]: E1008 09:29:52.854488 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a257281-2eda-4f01-920f-02757ef0fb37" containerName="cinder-scheduler" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.854494 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a257281-2eda-4f01-920f-02757ef0fb37" containerName="cinder-scheduler" Oct 08 09:29:52 crc kubenswrapper[4744]: E1008 09:29:52.854513 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8306420e-9e3d-48a6-87b9-af6c61f0d0ba" containerName="init" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.854519 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8306420e-9e3d-48a6-87b9-af6c61f0d0ba" containerName="init" Oct 08 09:29:52 crc kubenswrapper[4744]: E1008 09:29:52.854543 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a257281-2eda-4f01-920f-02757ef0fb37" containerName="probe" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.854551 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a257281-2eda-4f01-920f-02757ef0fb37" containerName="probe" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.854740 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a257281-2eda-4f01-920f-02757ef0fb37" containerName="probe" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.854759 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8306420e-9e3d-48a6-87b9-af6c61f0d0ba" containerName="dnsmasq-dns" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.854769 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a257281-2eda-4f01-920f-02757ef0fb37" containerName="cinder-scheduler" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.866848 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.889864 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.899117 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.912757 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bc72-account-create-gfwls" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.959999 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49177d36-177b-4b09-a93e-3d602ff30a27-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.960119 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49177d36-177b-4b09-a93e-3d602ff30a27-scripts\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.960237 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6m82\" (UniqueName: \"kubernetes.io/projected/49177d36-177b-4b09-a93e-3d602ff30a27-kube-api-access-k6m82\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.960278 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49177d36-177b-4b09-a93e-3d602ff30a27-config-data\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.960311 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49177d36-177b-4b09-a93e-3d602ff30a27-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:52 crc kubenswrapper[4744]: I1008 09:29:52.960333 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/49177d36-177b-4b09-a93e-3d602ff30a27-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.062220 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7c5j\" (UniqueName: \"kubernetes.io/projected/2edcd16b-76ad-4a92-b75b-033b9edd0024-kube-api-access-r7c5j\") pod \"2edcd16b-76ad-4a92-b75b-033b9edd0024\" (UID: \"2edcd16b-76ad-4a92-b75b-033b9edd0024\") " Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.062768 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49177d36-177b-4b09-a93e-3d602ff30a27-scripts\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.062845 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6m82\" (UniqueName: \"kubernetes.io/projected/49177d36-177b-4b09-a93e-3d602ff30a27-kube-api-access-k6m82\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.062881 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49177d36-177b-4b09-a93e-3d602ff30a27-config-data\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.062898 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49177d36-177b-4b09-a93e-3d602ff30a27-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.062918 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/49177d36-177b-4b09-a93e-3d602ff30a27-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.062985 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49177d36-177b-4b09-a93e-3d602ff30a27-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.063131 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/49177d36-177b-4b09-a93e-3d602ff30a27-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.070402 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49177d36-177b-4b09-a93e-3d602ff30a27-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.070708 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/49177d36-177b-4b09-a93e-3d602ff30a27-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.071204 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2edcd16b-76ad-4a92-b75b-033b9edd0024-kube-api-access-r7c5j" (OuterVolumeSpecName: "kube-api-access-r7c5j") pod "2edcd16b-76ad-4a92-b75b-033b9edd0024" (UID: "2edcd16b-76ad-4a92-b75b-033b9edd0024"). InnerVolumeSpecName "kube-api-access-r7c5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.072652 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/49177d36-177b-4b09-a93e-3d602ff30a27-scripts\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.072704 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49177d36-177b-4b09-a93e-3d602ff30a27-config-data\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.076710 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b88-account-create-v6x5w" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.087515 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6m82\" (UniqueName: \"kubernetes.io/projected/49177d36-177b-4b09-a93e-3d602ff30a27-kube-api-access-k6m82\") pod \"cinder-scheduler-0\" (UID: \"49177d36-177b-4b09-a93e-3d602ff30a27\") " pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.164535 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktmhs\" (UniqueName: \"kubernetes.io/projected/7edf84ea-4d5a-4764-b6a7-ea03f6e312e4-kube-api-access-ktmhs\") pod \"7edf84ea-4d5a-4764-b6a7-ea03f6e312e4\" (UID: \"7edf84ea-4d5a-4764-b6a7-ea03f6e312e4\") " Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.165136 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7c5j\" (UniqueName: \"kubernetes.io/projected/2edcd16b-76ad-4a92-b75b-033b9edd0024-kube-api-access-r7c5j\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.169524 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7edf84ea-4d5a-4764-b6a7-ea03f6e312e4-kube-api-access-ktmhs" (OuterVolumeSpecName: "kube-api-access-ktmhs") pod "7edf84ea-4d5a-4764-b6a7-ea03f6e312e4" (UID: "7edf84ea-4d5a-4764-b6a7-ea03f6e312e4"). InnerVolumeSpecName "kube-api-access-ktmhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.232079 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.267235 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktmhs\" (UniqueName: \"kubernetes.io/projected/7edf84ea-4d5a-4764-b6a7-ea03f6e312e4-kube-api-access-ktmhs\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.428087 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-bc72-account-create-gfwls" event={"ID":"2edcd16b-76ad-4a92-b75b-033b9edd0024","Type":"ContainerDied","Data":"d855e82220b5b6051eab40e717ea5cb287ad149342b83df5f6a118f2957be4f5"} Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.428141 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d855e82220b5b6051eab40e717ea5cb287ad149342b83df5f6a118f2957be4f5" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.428215 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-bc72-account-create-gfwls" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.435344 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b88-account-create-v6x5w" event={"ID":"7edf84ea-4d5a-4764-b6a7-ea03f6e312e4","Type":"ContainerDied","Data":"dc2bbe45ef6bf258035593e0805d687de532def819071ba206994a6075f76157"} Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.435397 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc2bbe45ef6bf258035593e0805d687de532def819071ba206994a6075f76157" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.435482 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b88-account-create-v6x5w" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.476848 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a257281-2eda-4f01-920f-02757ef0fb37" path="/var/lib/kubelet/pods/8a257281-2eda-4f01-920f-02757ef0fb37/volumes" Oct 08 09:29:53 crc kubenswrapper[4744]: I1008 09:29:53.799657 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.454316 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"49177d36-177b-4b09-a93e-3d602ff30a27","Type":"ContainerStarted","Data":"0ca0dbf118c622966a01b64f30eb4f66ebf18195fddb4affd8f0d626e9b63d52"} Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.457139 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d","Type":"ContainerStarted","Data":"8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50"} Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.457242 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="ceilometer-central-agent" containerID="cri-o://10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245" gracePeriod=30 Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.457287 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.457283 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="proxy-httpd" containerID="cri-o://8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50" gracePeriod=30 Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.457313 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="sg-core" containerID="cri-o://1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95" gracePeriod=30 Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.457342 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="ceilometer-notification-agent" containerID="cri-o://514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6" gracePeriod=30 Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.503171 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.737228275 podStartE2EDuration="6.503144357s" podCreationTimestamp="2025-10-08 09:29:48 +0000 UTC" firstStartedPulling="2025-10-08 09:29:49.482647347 +0000 UTC m=+1084.730292586" lastFinishedPulling="2025-10-08 09:29:53.248563429 +0000 UTC m=+1088.496208668" observedRunningTime="2025-10-08 09:29:54.489540914 +0000 UTC m=+1089.737186173" watchObservedRunningTime="2025-10-08 09:29:54.503144357 +0000 UTC m=+1089.750789596" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.580916 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4txff"] Oct 08 09:29:54 crc kubenswrapper[4744]: E1008 09:29:54.583122 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2edcd16b-76ad-4a92-b75b-033b9edd0024" containerName="mariadb-account-create" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.583144 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2edcd16b-76ad-4a92-b75b-033b9edd0024" containerName="mariadb-account-create" Oct 08 09:29:54 crc kubenswrapper[4744]: E1008 09:29:54.583612 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7edf84ea-4d5a-4764-b6a7-ea03f6e312e4" containerName="mariadb-account-create" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.583624 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7edf84ea-4d5a-4764-b6a7-ea03f6e312e4" containerName="mariadb-account-create" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.583976 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2edcd16b-76ad-4a92-b75b-033b9edd0024" containerName="mariadb-account-create" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.584004 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7edf84ea-4d5a-4764-b6a7-ea03f6e312e4" containerName="mariadb-account-create" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.585185 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.611313 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.613030 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-szg46" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.613113 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.630891 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4txff"] Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.714117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-config-data\") pod \"nova-cell0-conductor-db-sync-4txff\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.714188 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-scripts\") pod \"nova-cell0-conductor-db-sync-4txff\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.714242 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-4txff\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.714277 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt254\" (UniqueName: \"kubernetes.io/projected/93968d7d-7816-4580-8a04-f4ccacd584b6-kube-api-access-tt254\") pod \"nova-cell0-conductor-db-sync-4txff\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.730837 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.821665 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt254\" (UniqueName: \"kubernetes.io/projected/93968d7d-7816-4580-8a04-f4ccacd584b6-kube-api-access-tt254\") pod \"nova-cell0-conductor-db-sync-4txff\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.825555 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-config-data\") pod \"nova-cell0-conductor-db-sync-4txff\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.826128 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-scripts\") pod \"nova-cell0-conductor-db-sync-4txff\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.827182 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-4txff\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.838006 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-4txff\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.842811 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-config-data\") pod \"nova-cell0-conductor-db-sync-4txff\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.860543 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-scripts\") pod \"nova-cell0-conductor-db-sync-4txff\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.896101 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt254\" (UniqueName: \"kubernetes.io/projected/93968d7d-7816-4580-8a04-f4ccacd584b6-kube-api-access-tt254\") pod \"nova-cell0-conductor-db-sync-4txff\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:54 crc kubenswrapper[4744]: I1008 09:29:54.934847 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:29:55 crc kubenswrapper[4744]: E1008 09:29:55.263892 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5abb00d_f75c_4cb1_abf3_6c38e7957b1d.slice/crio-8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5abb00d_f75c_4cb1_abf3_6c38e7957b1d.slice/crio-conmon-8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50.scope\": RecentStats: unable to find data in memory cache]" Oct 08 09:29:55 crc kubenswrapper[4744]: I1008 09:29:55.533021 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"49177d36-177b-4b09-a93e-3d602ff30a27","Type":"ContainerStarted","Data":"6993bbe0a1b1be30cf9009b0e36d278eec8ccf61abe6a1d16a9c4c021d3f5e74"} Oct 08 09:29:55 crc kubenswrapper[4744]: I1008 09:29:55.539525 4744 generic.go:334] "Generic (PLEG): container finished" podID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerID="8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50" exitCode=0 Oct 08 09:29:55 crc kubenswrapper[4744]: I1008 09:29:55.539581 4744 generic.go:334] "Generic (PLEG): container finished" podID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerID="1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95" exitCode=2 Oct 08 09:29:55 crc kubenswrapper[4744]: I1008 09:29:55.539592 4744 generic.go:334] "Generic (PLEG): container finished" podID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerID="514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6" exitCode=0 Oct 08 09:29:55 crc kubenswrapper[4744]: I1008 09:29:55.539622 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d","Type":"ContainerDied","Data":"8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50"} Oct 08 09:29:55 crc kubenswrapper[4744]: I1008 09:29:55.539676 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d","Type":"ContainerDied","Data":"1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95"} Oct 08 09:29:55 crc kubenswrapper[4744]: I1008 09:29:55.539688 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d","Type":"ContainerDied","Data":"514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6"} Oct 08 09:29:55 crc kubenswrapper[4744]: I1008 09:29:55.708401 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4txff"] Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.073350 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.182220 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-config-data\") pod \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.182290 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-scripts\") pod \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.182322 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-combined-ca-bundle\") pod \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.182423 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-sg-core-conf-yaml\") pod \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.182476 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5gpk\" (UniqueName: \"kubernetes.io/projected/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-kube-api-access-h5gpk\") pod \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.182517 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-run-httpd\") pod \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.182634 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-log-httpd\") pod \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\" (UID: \"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d\") " Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.186040 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" (UID: "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.188475 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" (UID: "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.215701 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-kube-api-access-h5gpk" (OuterVolumeSpecName: "kube-api-access-h5gpk") pod "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" (UID: "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d"). InnerVolumeSpecName "kube-api-access-h5gpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.215835 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-scripts" (OuterVolumeSpecName: "scripts") pod "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" (UID: "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.280656 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" (UID: "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.284916 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.284950 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.284960 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.284972 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5gpk\" (UniqueName: \"kubernetes.io/projected/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-kube-api-access-h5gpk\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.284986 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.342747 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" (UID: "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.389803 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.400482 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-config-data" (OuterVolumeSpecName: "config-data") pod "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" (UID: "a5abb00d-f75c-4cb1-abf3-6c38e7957b1d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.491140 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.561784 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4txff" event={"ID":"93968d7d-7816-4580-8a04-f4ccacd584b6","Type":"ContainerStarted","Data":"7eb8561cd2f6878e0ed559766b692a2af187855fd1c06e31b358d7492d933022"} Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.578621 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"49177d36-177b-4b09-a93e-3d602ff30a27","Type":"ContainerStarted","Data":"d1bad87717c25254b9990e267f553ae5ccd03f900c7ca0f113eabf08bf23419d"} Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.606618 4744 generic.go:334] "Generic (PLEG): container finished" podID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerID="10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245" exitCode=0 Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.606674 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d","Type":"ContainerDied","Data":"10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245"} Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.606710 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5abb00d-f75c-4cb1-abf3-6c38e7957b1d","Type":"ContainerDied","Data":"c1dfe6c2358b18d1d80ad4a53c669a6f1526d3cc838e4dced863870b58a27000"} Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.606731 4744 scope.go:117] "RemoveContainer" containerID="8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.606879 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.632684 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.632648294 podStartE2EDuration="4.632648294s" podCreationTimestamp="2025-10-08 09:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:29:56.617252013 +0000 UTC m=+1091.864897252" watchObservedRunningTime="2025-10-08 09:29:56.632648294 +0000 UTC m=+1091.880293543" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.678782 4744 scope.go:117] "RemoveContainer" containerID="1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.711472 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.731651 4744 scope.go:117] "RemoveContainer" containerID="514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.740501 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.786476 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:56 crc kubenswrapper[4744]: E1008 09:29:56.787053 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="sg-core" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.787072 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="sg-core" Oct 08 09:29:56 crc kubenswrapper[4744]: E1008 09:29:56.787087 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="ceilometer-notification-agent" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.787094 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="ceilometer-notification-agent" Oct 08 09:29:56 crc kubenswrapper[4744]: E1008 09:29:56.787109 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="proxy-httpd" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.787115 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="proxy-httpd" Oct 08 09:29:56 crc kubenswrapper[4744]: E1008 09:29:56.787129 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="ceilometer-central-agent" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.787135 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="ceilometer-central-agent" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.787320 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="proxy-httpd" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.787333 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="ceilometer-notification-agent" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.787341 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="ceilometer-central-agent" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.787359 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" containerName="sg-core" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.789056 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.794263 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.794581 4744 scope.go:117] "RemoveContainer" containerID="10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.797853 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.802263 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.931104 4744 scope.go:117] "RemoveContainer" containerID="8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50" Oct 08 09:29:56 crc kubenswrapper[4744]: E1008 09:29:56.932212 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50\": container with ID starting with 8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50 not found: ID does not exist" containerID="8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.932262 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50"} err="failed to get container status \"8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50\": rpc error: code = NotFound desc = could not find container \"8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50\": container with ID starting with 8569e28a35be864debd41bce89540daeddd8fd1c1baf68b53eb9c22470f3ba50 not found: ID does not exist" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.932303 4744 scope.go:117] "RemoveContainer" containerID="1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95" Oct 08 09:29:56 crc kubenswrapper[4744]: E1008 09:29:56.932809 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95\": container with ID starting with 1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95 not found: ID does not exist" containerID="1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.932838 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95"} err="failed to get container status \"1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95\": rpc error: code = NotFound desc = could not find container \"1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95\": container with ID starting with 1b06c775b85547e47519edccae148b5a9ed19108fe49952d00e99e5b68afee95 not found: ID does not exist" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.932858 4744 scope.go:117] "RemoveContainer" containerID="514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6" Oct 08 09:29:56 crc kubenswrapper[4744]: E1008 09:29:56.933044 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6\": container with ID starting with 514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6 not found: ID does not exist" containerID="514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.933066 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6"} err="failed to get container status \"514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6\": rpc error: code = NotFound desc = could not find container \"514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6\": container with ID starting with 514788f4f0d8fce9d6bc5f005d1d8d3827f5a33987b82e3878617b6b2f8822f6 not found: ID does not exist" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.933079 4744 scope.go:117] "RemoveContainer" containerID="10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245" Oct 08 09:29:56 crc kubenswrapper[4744]: E1008 09:29:56.933756 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245\": container with ID starting with 10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245 not found: ID does not exist" containerID="10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.933816 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245"} err="failed to get container status \"10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245\": rpc error: code = NotFound desc = could not find container \"10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245\": container with ID starting with 10b999929790cdce93e5b35a87036ae18cb89dd1255d0ee876952543d8ff1245 not found: ID does not exist" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.939942 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.940015 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-config-data\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.940048 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.940114 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-scripts\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.940131 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-log-httpd\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.940168 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwmf5\" (UniqueName: \"kubernetes.io/projected/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-kube-api-access-xwmf5\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:56 crc kubenswrapper[4744]: I1008 09:29:56.940194 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-run-httpd\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.043239 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.043323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-config-data\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.043408 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.043503 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-scripts\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.043527 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-log-httpd\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.043590 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwmf5\" (UniqueName: \"kubernetes.io/projected/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-kube-api-access-xwmf5\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.043635 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-run-httpd\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.044104 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-run-httpd\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.045525 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-log-httpd\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.050422 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.063925 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-config-data\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.068788 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-scripts\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.078135 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.099279 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwmf5\" (UniqueName: \"kubernetes.io/projected/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-kube-api-access-xwmf5\") pod \"ceilometer-0\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.126042 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.483028 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5abb00d-f75c-4cb1-abf3-6c38e7957b1d" path="/var/lib/kubelet/pods/a5abb00d-f75c-4cb1-abf3-6c38e7957b1d/volumes" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.555069 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.648619 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f88fa333-9e1b-4ccb-a708-a256cd4f64e2","Type":"ContainerStarted","Data":"36af241555aa16e87efe091adba35e1e544c3e34ee32167b703409361e36b48d"} Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.732798 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6b4cb6945c-kjx6m" Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.820674 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7f466fb8c4-5nw45"] Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.820909 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7f466fb8c4-5nw45" podUID="94715e15-b083-4428-95d6-d7e66990af08" containerName="neutron-api" containerID="cri-o://dbafe4df8d4eb5e4343d7baac824d5f64992310c6264c9583d609941e07ebfd7" gracePeriod=30 Oct 08 09:29:57 crc kubenswrapper[4744]: I1008 09:29:57.821352 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7f466fb8c4-5nw45" podUID="94715e15-b083-4428-95d6-d7e66990af08" containerName="neutron-httpd" containerID="cri-o://cd4ce2223ffbac03d43a8d5f76019aaa87e93027c60ead4fd2f9caaa285e42cb" gracePeriod=30 Oct 08 09:29:58 crc kubenswrapper[4744]: I1008 09:29:58.232748 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 09:29:58 crc kubenswrapper[4744]: I1008 09:29:58.675237 4744 generic.go:334] "Generic (PLEG): container finished" podID="94715e15-b083-4428-95d6-d7e66990af08" containerID="cd4ce2223ffbac03d43a8d5f76019aaa87e93027c60ead4fd2f9caaa285e42cb" exitCode=0 Oct 08 09:29:58 crc kubenswrapper[4744]: I1008 09:29:58.675662 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f466fb8c4-5nw45" event={"ID":"94715e15-b083-4428-95d6-d7e66990af08","Type":"ContainerDied","Data":"cd4ce2223ffbac03d43a8d5f76019aaa87e93027c60ead4fd2f9caaa285e42cb"} Oct 08 09:29:58 crc kubenswrapper[4744]: I1008 09:29:58.692766 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f88fa333-9e1b-4ccb-a708-a256cd4f64e2","Type":"ContainerStarted","Data":"53b1f69045cb9ec2b011e77c8bc9617e5dcf8a4b372b70835b35dd8861a64768"} Oct 08 09:29:59 crc kubenswrapper[4744]: I1008 09:29:59.126542 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 08 09:29:59 crc kubenswrapper[4744]: I1008 09:29:59.283162 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-66f4-account-create-5dh4v"] Oct 08 09:29:59 crc kubenswrapper[4744]: I1008 09:29:59.284753 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-66f4-account-create-5dh4v" Oct 08 09:29:59 crc kubenswrapper[4744]: I1008 09:29:59.292942 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 08 09:29:59 crc kubenswrapper[4744]: I1008 09:29:59.304739 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-66f4-account-create-5dh4v"] Oct 08 09:29:59 crc kubenswrapper[4744]: I1008 09:29:59.411148 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsjrc\" (UniqueName: \"kubernetes.io/projected/800eea78-e2f9-40d8-a2f9-bfe60b9adea9-kube-api-access-lsjrc\") pod \"nova-cell1-66f4-account-create-5dh4v\" (UID: \"800eea78-e2f9-40d8-a2f9-bfe60b9adea9\") " pod="openstack/nova-cell1-66f4-account-create-5dh4v" Oct 08 09:29:59 crc kubenswrapper[4744]: I1008 09:29:59.514129 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsjrc\" (UniqueName: \"kubernetes.io/projected/800eea78-e2f9-40d8-a2f9-bfe60b9adea9-kube-api-access-lsjrc\") pod \"nova-cell1-66f4-account-create-5dh4v\" (UID: \"800eea78-e2f9-40d8-a2f9-bfe60b9adea9\") " pod="openstack/nova-cell1-66f4-account-create-5dh4v" Oct 08 09:29:59 crc kubenswrapper[4744]: I1008 09:29:59.549198 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsjrc\" (UniqueName: \"kubernetes.io/projected/800eea78-e2f9-40d8-a2f9-bfe60b9adea9-kube-api-access-lsjrc\") pod \"nova-cell1-66f4-account-create-5dh4v\" (UID: \"800eea78-e2f9-40d8-a2f9-bfe60b9adea9\") " pod="openstack/nova-cell1-66f4-account-create-5dh4v" Oct 08 09:29:59 crc kubenswrapper[4744]: I1008 09:29:59.609853 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-66f4-account-create-5dh4v" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.165479 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5"] Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.168047 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.172572 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.172754 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.175403 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5"] Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.320740 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-66f4-account-create-5dh4v"] Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.332830 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/162601ad-6234-499e-acfe-6e8fba25d8a0-config-volume\") pod \"collect-profiles-29331930-zgwg5\" (UID: \"162601ad-6234-499e-acfe-6e8fba25d8a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.332937 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/162601ad-6234-499e-acfe-6e8fba25d8a0-secret-volume\") pod \"collect-profiles-29331930-zgwg5\" (UID: \"162601ad-6234-499e-acfe-6e8fba25d8a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.332998 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw8w8\" (UniqueName: \"kubernetes.io/projected/162601ad-6234-499e-acfe-6e8fba25d8a0-kube-api-access-xw8w8\") pod \"collect-profiles-29331930-zgwg5\" (UID: \"162601ad-6234-499e-acfe-6e8fba25d8a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.442642 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/162601ad-6234-499e-acfe-6e8fba25d8a0-config-volume\") pod \"collect-profiles-29331930-zgwg5\" (UID: \"162601ad-6234-499e-acfe-6e8fba25d8a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.442758 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/162601ad-6234-499e-acfe-6e8fba25d8a0-secret-volume\") pod \"collect-profiles-29331930-zgwg5\" (UID: \"162601ad-6234-499e-acfe-6e8fba25d8a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.442823 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw8w8\" (UniqueName: \"kubernetes.io/projected/162601ad-6234-499e-acfe-6e8fba25d8a0-kube-api-access-xw8w8\") pod \"collect-profiles-29331930-zgwg5\" (UID: \"162601ad-6234-499e-acfe-6e8fba25d8a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.448054 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/162601ad-6234-499e-acfe-6e8fba25d8a0-config-volume\") pod \"collect-profiles-29331930-zgwg5\" (UID: \"162601ad-6234-499e-acfe-6e8fba25d8a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.474739 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw8w8\" (UniqueName: \"kubernetes.io/projected/162601ad-6234-499e-acfe-6e8fba25d8a0-kube-api-access-xw8w8\") pod \"collect-profiles-29331930-zgwg5\" (UID: \"162601ad-6234-499e-acfe-6e8fba25d8a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.487246 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/162601ad-6234-499e-acfe-6e8fba25d8a0-secret-volume\") pod \"collect-profiles-29331930-zgwg5\" (UID: \"162601ad-6234-499e-acfe-6e8fba25d8a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.511429 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.823566 4744 generic.go:334] "Generic (PLEG): container finished" podID="800eea78-e2f9-40d8-a2f9-bfe60b9adea9" containerID="5d4430a6d79b848025ab9dc8f42b5fafa7e4cf074ebf7e7b5ce0acec938fecd9" exitCode=0 Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.824333 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-66f4-account-create-5dh4v" event={"ID":"800eea78-e2f9-40d8-a2f9-bfe60b9adea9","Type":"ContainerDied","Data":"5d4430a6d79b848025ab9dc8f42b5fafa7e4cf074ebf7e7b5ce0acec938fecd9"} Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.824515 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-66f4-account-create-5dh4v" event={"ID":"800eea78-e2f9-40d8-a2f9-bfe60b9adea9","Type":"ContainerStarted","Data":"84140f3bd1c35d3115c341a803b2316e178df5f33188e251bd16c964ae7e44d5"} Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.838322 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f88fa333-9e1b-4ccb-a708-a256cd4f64e2","Type":"ContainerStarted","Data":"64828286519fbe58cee05ad6bce0d017e0f80d7557d117f4eee2dcfe9e5d542d"} Oct 08 09:30:00 crc kubenswrapper[4744]: W1008 09:30:00.980640 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod162601ad_6234_499e_acfe_6e8fba25d8a0.slice/crio-a4ac7afa19977fefca716f90c26278150dc1dbd7d9106915ae47b031f237f2db WatchSource:0}: Error finding container a4ac7afa19977fefca716f90c26278150dc1dbd7d9106915ae47b031f237f2db: Status 404 returned error can't find the container with id a4ac7afa19977fefca716f90c26278150dc1dbd7d9106915ae47b031f237f2db Oct 08 09:30:00 crc kubenswrapper[4744]: I1008 09:30:00.985055 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5"] Oct 08 09:30:01 crc kubenswrapper[4744]: I1008 09:30:01.715046 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:01 crc kubenswrapper[4744]: I1008 09:30:01.851844 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" event={"ID":"162601ad-6234-499e-acfe-6e8fba25d8a0","Type":"ContainerStarted","Data":"d6dede98da6982a5437e6ceaf32ddbaac0e3cc9e6c3119b070f5829ec0c192b6"} Oct 08 09:30:01 crc kubenswrapper[4744]: I1008 09:30:01.851891 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" event={"ID":"162601ad-6234-499e-acfe-6e8fba25d8a0","Type":"ContainerStarted","Data":"a4ac7afa19977fefca716f90c26278150dc1dbd7d9106915ae47b031f237f2db"} Oct 08 09:30:01 crc kubenswrapper[4744]: I1008 09:30:01.859188 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f88fa333-9e1b-4ccb-a708-a256cd4f64e2","Type":"ContainerStarted","Data":"f636e30c3e503c24bdba4a6e528690233c9cc5fcd7ccb3787176cb8837f823ab"} Oct 08 09:30:01 crc kubenswrapper[4744]: I1008 09:30:01.875865 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" podStartSLOduration=1.875843868 podStartE2EDuration="1.875843868s" podCreationTimestamp="2025-10-08 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:30:01.875697654 +0000 UTC m=+1097.123342893" watchObservedRunningTime="2025-10-08 09:30:01.875843868 +0000 UTC m=+1097.123489107" Oct 08 09:30:02 crc kubenswrapper[4744]: I1008 09:30:02.301538 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-66f4-account-create-5dh4v" Oct 08 09:30:02 crc kubenswrapper[4744]: I1008 09:30:02.388939 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsjrc\" (UniqueName: \"kubernetes.io/projected/800eea78-e2f9-40d8-a2f9-bfe60b9adea9-kube-api-access-lsjrc\") pod \"800eea78-e2f9-40d8-a2f9-bfe60b9adea9\" (UID: \"800eea78-e2f9-40d8-a2f9-bfe60b9adea9\") " Oct 08 09:30:02 crc kubenswrapper[4744]: I1008 09:30:02.394620 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/800eea78-e2f9-40d8-a2f9-bfe60b9adea9-kube-api-access-lsjrc" (OuterVolumeSpecName: "kube-api-access-lsjrc") pod "800eea78-e2f9-40d8-a2f9-bfe60b9adea9" (UID: "800eea78-e2f9-40d8-a2f9-bfe60b9adea9"). InnerVolumeSpecName "kube-api-access-lsjrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:02 crc kubenswrapper[4744]: I1008 09:30:02.490916 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsjrc\" (UniqueName: \"kubernetes.io/projected/800eea78-e2f9-40d8-a2f9-bfe60b9adea9-kube-api-access-lsjrc\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:02 crc kubenswrapper[4744]: I1008 09:30:02.881473 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-66f4-account-create-5dh4v" Oct 08 09:30:02 crc kubenswrapper[4744]: I1008 09:30:02.881499 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-66f4-account-create-5dh4v" event={"ID":"800eea78-e2f9-40d8-a2f9-bfe60b9adea9","Type":"ContainerDied","Data":"84140f3bd1c35d3115c341a803b2316e178df5f33188e251bd16c964ae7e44d5"} Oct 08 09:30:02 crc kubenswrapper[4744]: I1008 09:30:02.881562 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84140f3bd1c35d3115c341a803b2316e178df5f33188e251bd16c964ae7e44d5" Oct 08 09:30:02 crc kubenswrapper[4744]: I1008 09:30:02.886023 4744 generic.go:334] "Generic (PLEG): container finished" podID="162601ad-6234-499e-acfe-6e8fba25d8a0" containerID="d6dede98da6982a5437e6ceaf32ddbaac0e3cc9e6c3119b070f5829ec0c192b6" exitCode=0 Oct 08 09:30:02 crc kubenswrapper[4744]: I1008 09:30:02.886069 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" event={"ID":"162601ad-6234-499e-acfe-6e8fba25d8a0","Type":"ContainerDied","Data":"d6dede98da6982a5437e6ceaf32ddbaac0e3cc9e6c3119b070f5829ec0c192b6"} Oct 08 09:30:03 crc kubenswrapper[4744]: I1008 09:30:03.525473 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 09:30:03 crc kubenswrapper[4744]: I1008 09:30:03.912176 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f88fa333-9e1b-4ccb-a708-a256cd4f64e2","Type":"ContainerStarted","Data":"c9bf093a88029cebd2bb4fc6de42b9a0f69ed5b23b0d1cbdc0f5c781cd61e568"} Oct 08 09:30:03 crc kubenswrapper[4744]: I1008 09:30:03.913324 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 09:30:03 crc kubenswrapper[4744]: I1008 09:30:03.912581 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="ceilometer-notification-agent" containerID="cri-o://64828286519fbe58cee05ad6bce0d017e0f80d7557d117f4eee2dcfe9e5d542d" gracePeriod=30 Oct 08 09:30:03 crc kubenswrapper[4744]: I1008 09:30:03.912516 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="proxy-httpd" containerID="cri-o://c9bf093a88029cebd2bb4fc6de42b9a0f69ed5b23b0d1cbdc0f5c781cd61e568" gracePeriod=30 Oct 08 09:30:03 crc kubenswrapper[4744]: I1008 09:30:03.912514 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="sg-core" containerID="cri-o://f636e30c3e503c24bdba4a6e528690233c9cc5fcd7ccb3787176cb8837f823ab" gracePeriod=30 Oct 08 09:30:03 crc kubenswrapper[4744]: I1008 09:30:03.912448 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="ceilometer-central-agent" containerID="cri-o://53b1f69045cb9ec2b011e77c8bc9617e5dcf8a4b372b70835b35dd8861a64768" gracePeriod=30 Oct 08 09:30:03 crc kubenswrapper[4744]: I1008 09:30:03.941616 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.840477425 podStartE2EDuration="7.941590471s" podCreationTimestamp="2025-10-08 09:29:56 +0000 UTC" firstStartedPulling="2025-10-08 09:29:57.574082411 +0000 UTC m=+1092.821727650" lastFinishedPulling="2025-10-08 09:30:02.675195457 +0000 UTC m=+1097.922840696" observedRunningTime="2025-10-08 09:30:03.937781089 +0000 UTC m=+1099.185426348" watchObservedRunningTime="2025-10-08 09:30:03.941590471 +0000 UTC m=+1099.189235710" Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.531675 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.650246 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/162601ad-6234-499e-acfe-6e8fba25d8a0-config-volume\") pod \"162601ad-6234-499e-acfe-6e8fba25d8a0\" (UID: \"162601ad-6234-499e-acfe-6e8fba25d8a0\") " Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.651695 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/162601ad-6234-499e-acfe-6e8fba25d8a0-secret-volume\") pod \"162601ad-6234-499e-acfe-6e8fba25d8a0\" (UID: \"162601ad-6234-499e-acfe-6e8fba25d8a0\") " Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.651745 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw8w8\" (UniqueName: \"kubernetes.io/projected/162601ad-6234-499e-acfe-6e8fba25d8a0-kube-api-access-xw8w8\") pod \"162601ad-6234-499e-acfe-6e8fba25d8a0\" (UID: \"162601ad-6234-499e-acfe-6e8fba25d8a0\") " Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.651073 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/162601ad-6234-499e-acfe-6e8fba25d8a0-config-volume" (OuterVolumeSpecName: "config-volume") pod "162601ad-6234-499e-acfe-6e8fba25d8a0" (UID: "162601ad-6234-499e-acfe-6e8fba25d8a0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.659624 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/162601ad-6234-499e-acfe-6e8fba25d8a0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "162601ad-6234-499e-acfe-6e8fba25d8a0" (UID: "162601ad-6234-499e-acfe-6e8fba25d8a0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.659916 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/162601ad-6234-499e-acfe-6e8fba25d8a0-kube-api-access-xw8w8" (OuterVolumeSpecName: "kube-api-access-xw8w8") pod "162601ad-6234-499e-acfe-6e8fba25d8a0" (UID: "162601ad-6234-499e-acfe-6e8fba25d8a0"). InnerVolumeSpecName "kube-api-access-xw8w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.753720 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw8w8\" (UniqueName: \"kubernetes.io/projected/162601ad-6234-499e-acfe-6e8fba25d8a0-kube-api-access-xw8w8\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.753757 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/162601ad-6234-499e-acfe-6e8fba25d8a0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.753767 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/162601ad-6234-499e-acfe-6e8fba25d8a0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.963202 4744 generic.go:334] "Generic (PLEG): container finished" podID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerID="c9bf093a88029cebd2bb4fc6de42b9a0f69ed5b23b0d1cbdc0f5c781cd61e568" exitCode=0 Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.963489 4744 generic.go:334] "Generic (PLEG): container finished" podID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerID="f636e30c3e503c24bdba4a6e528690233c9cc5fcd7ccb3787176cb8837f823ab" exitCode=2 Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.963274 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f88fa333-9e1b-4ccb-a708-a256cd4f64e2","Type":"ContainerDied","Data":"c9bf093a88029cebd2bb4fc6de42b9a0f69ed5b23b0d1cbdc0f5c781cd61e568"} Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.963530 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f88fa333-9e1b-4ccb-a708-a256cd4f64e2","Type":"ContainerDied","Data":"f636e30c3e503c24bdba4a6e528690233c9cc5fcd7ccb3787176cb8837f823ab"} Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.963543 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f88fa333-9e1b-4ccb-a708-a256cd4f64e2","Type":"ContainerDied","Data":"64828286519fbe58cee05ad6bce0d017e0f80d7557d117f4eee2dcfe9e5d542d"} Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.963497 4744 generic.go:334] "Generic (PLEG): container finished" podID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerID="64828286519fbe58cee05ad6bce0d017e0f80d7557d117f4eee2dcfe9e5d542d" exitCode=0 Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.963559 4744 generic.go:334] "Generic (PLEG): container finished" podID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerID="53b1f69045cb9ec2b011e77c8bc9617e5dcf8a4b372b70835b35dd8861a64768" exitCode=0 Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.963624 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f88fa333-9e1b-4ccb-a708-a256cd4f64e2","Type":"ContainerDied","Data":"53b1f69045cb9ec2b011e77c8bc9617e5dcf8a4b372b70835b35dd8861a64768"} Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.965587 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" event={"ID":"162601ad-6234-499e-acfe-6e8fba25d8a0","Type":"ContainerDied","Data":"a4ac7afa19977fefca716f90c26278150dc1dbd7d9106915ae47b031f237f2db"} Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.965618 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4ac7afa19977fefca716f90c26278150dc1dbd7d9106915ae47b031f237f2db" Oct 08 09:30:04 crc kubenswrapper[4744]: I1008 09:30:04.965667 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5" Oct 08 09:30:05 crc kubenswrapper[4744]: I1008 09:30:05.985138 4744 generic.go:334] "Generic (PLEG): container finished" podID="94715e15-b083-4428-95d6-d7e66990af08" containerID="dbafe4df8d4eb5e4343d7baac824d5f64992310c6264c9583d609941e07ebfd7" exitCode=0 Oct 08 09:30:05 crc kubenswrapper[4744]: I1008 09:30:05.985485 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f466fb8c4-5nw45" event={"ID":"94715e15-b083-4428-95d6-d7e66990af08","Type":"ContainerDied","Data":"dbafe4df8d4eb5e4343d7baac824d5f64992310c6264c9583d609941e07ebfd7"} Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.531666 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.662410 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-scripts\") pod \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.662569 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-config-data\") pod \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.662646 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-log-httpd\") pod \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.662728 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-combined-ca-bundle\") pod \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.662838 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-sg-core-conf-yaml\") pod \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.662870 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwmf5\" (UniqueName: \"kubernetes.io/projected/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-kube-api-access-xwmf5\") pod \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.662895 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-run-httpd\") pod \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\" (UID: \"f88fa333-9e1b-4ccb-a708-a256cd4f64e2\") " Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.666688 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f88fa333-9e1b-4ccb-a708-a256cd4f64e2" (UID: "f88fa333-9e1b-4ccb-a708-a256cd4f64e2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.667929 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f88fa333-9e1b-4ccb-a708-a256cd4f64e2" (UID: "f88fa333-9e1b-4ccb-a708-a256cd4f64e2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.668256 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.668274 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.680979 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-kube-api-access-xwmf5" (OuterVolumeSpecName: "kube-api-access-xwmf5") pod "f88fa333-9e1b-4ccb-a708-a256cd4f64e2" (UID: "f88fa333-9e1b-4ccb-a708-a256cd4f64e2"). InnerVolumeSpecName "kube-api-access-xwmf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.691037 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-scripts" (OuterVolumeSpecName: "scripts") pod "f88fa333-9e1b-4ccb-a708-a256cd4f64e2" (UID: "f88fa333-9e1b-4ccb-a708-a256cd4f64e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.763311 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f88fa333-9e1b-4ccb-a708-a256cd4f64e2" (UID: "f88fa333-9e1b-4ccb-a708-a256cd4f64e2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.771965 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.771996 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.772005 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwmf5\" (UniqueName: \"kubernetes.io/projected/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-kube-api-access-xwmf5\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.838583 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f88fa333-9e1b-4ccb-a708-a256cd4f64e2" (UID: "f88fa333-9e1b-4ccb-a708-a256cd4f64e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.873260 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.893531 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-config-data" (OuterVolumeSpecName: "config-data") pod "f88fa333-9e1b-4ccb-a708-a256cd4f64e2" (UID: "f88fa333-9e1b-4ccb-a708-a256cd4f64e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:09 crc kubenswrapper[4744]: I1008 09:30:09.974730 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f88fa333-9e1b-4ccb-a708-a256cd4f64e2-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.035052 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f88fa333-9e1b-4ccb-a708-a256cd4f64e2","Type":"ContainerDied","Data":"36af241555aa16e87efe091adba35e1e544c3e34ee32167b703409361e36b48d"} Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.035127 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.035161 4744 scope.go:117] "RemoveContainer" containerID="c9bf093a88029cebd2bb4fc6de42b9a0f69ed5b23b0d1cbdc0f5c781cd61e568" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.091153 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.094427 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.104351 4744 scope.go:117] "RemoveContainer" containerID="f636e30c3e503c24bdba4a6e528690233c9cc5fcd7ccb3787176cb8837f823ab" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.109574 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.128006 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:10 crc kubenswrapper[4744]: E1008 09:30:10.128737 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94715e15-b083-4428-95d6-d7e66990af08" containerName="neutron-httpd" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.128758 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="94715e15-b083-4428-95d6-d7e66990af08" containerName="neutron-httpd" Oct 08 09:30:10 crc kubenswrapper[4744]: E1008 09:30:10.128771 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="ceilometer-notification-agent" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.128781 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="ceilometer-notification-agent" Oct 08 09:30:10 crc kubenswrapper[4744]: E1008 09:30:10.128790 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="162601ad-6234-499e-acfe-6e8fba25d8a0" containerName="collect-profiles" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.128798 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="162601ad-6234-499e-acfe-6e8fba25d8a0" containerName="collect-profiles" Oct 08 09:30:10 crc kubenswrapper[4744]: E1008 09:30:10.128807 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="sg-core" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.128813 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="sg-core" Oct 08 09:30:10 crc kubenswrapper[4744]: E1008 09:30:10.128835 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="800eea78-e2f9-40d8-a2f9-bfe60b9adea9" containerName="mariadb-account-create" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.128841 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="800eea78-e2f9-40d8-a2f9-bfe60b9adea9" containerName="mariadb-account-create" Oct 08 09:30:10 crc kubenswrapper[4744]: E1008 09:30:10.128858 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="ceilometer-central-agent" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.128866 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="ceilometer-central-agent" Oct 08 09:30:10 crc kubenswrapper[4744]: E1008 09:30:10.128882 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="proxy-httpd" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.128888 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="proxy-httpd" Oct 08 09:30:10 crc kubenswrapper[4744]: E1008 09:30:10.128910 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94715e15-b083-4428-95d6-d7e66990af08" containerName="neutron-api" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.128916 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="94715e15-b083-4428-95d6-d7e66990af08" containerName="neutron-api" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.129084 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="94715e15-b083-4428-95d6-d7e66990af08" containerName="neutron-api" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.129102 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="proxy-httpd" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.129119 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="162601ad-6234-499e-acfe-6e8fba25d8a0" containerName="collect-profiles" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.129130 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="sg-core" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.129147 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="94715e15-b083-4428-95d6-d7e66990af08" containerName="neutron-httpd" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.129158 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="ceilometer-notification-agent" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.129169 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="800eea78-e2f9-40d8-a2f9-bfe60b9adea9" containerName="mariadb-account-create" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.129180 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" containerName="ceilometer-central-agent" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.130801 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.134605 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.134775 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.144318 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.176784 4744 scope.go:117] "RemoveContainer" containerID="64828286519fbe58cee05ad6bce0d017e0f80d7557d117f4eee2dcfe9e5d542d" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.177530 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-combined-ca-bundle\") pod \"94715e15-b083-4428-95d6-d7e66990af08\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.177713 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-ovndb-tls-certs\") pod \"94715e15-b083-4428-95d6-d7e66990af08\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.177849 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpjjq\" (UniqueName: \"kubernetes.io/projected/94715e15-b083-4428-95d6-d7e66990af08-kube-api-access-gpjjq\") pod \"94715e15-b083-4428-95d6-d7e66990af08\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.177925 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-config\") pod \"94715e15-b083-4428-95d6-d7e66990af08\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.177984 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-httpd-config\") pod \"94715e15-b083-4428-95d6-d7e66990af08\" (UID: \"94715e15-b083-4428-95d6-d7e66990af08\") " Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.179321 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr26x\" (UniqueName: \"kubernetes.io/projected/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-kube-api-access-qr26x\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.179406 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-config-data\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.179491 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-log-httpd\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.179537 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.179598 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-scripts\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.179695 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-run-httpd\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.179738 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.184650 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "94715e15-b083-4428-95d6-d7e66990af08" (UID: "94715e15-b083-4428-95d6-d7e66990af08"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.198824 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94715e15-b083-4428-95d6-d7e66990af08-kube-api-access-gpjjq" (OuterVolumeSpecName: "kube-api-access-gpjjq") pod "94715e15-b083-4428-95d6-d7e66990af08" (UID: "94715e15-b083-4428-95d6-d7e66990af08"). InnerVolumeSpecName "kube-api-access-gpjjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.232800 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94715e15-b083-4428-95d6-d7e66990af08" (UID: "94715e15-b083-4428-95d6-d7e66990af08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.234619 4744 scope.go:117] "RemoveContainer" containerID="53b1f69045cb9ec2b011e77c8bc9617e5dcf8a4b372b70835b35dd8861a64768" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.270622 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-config" (OuterVolumeSpecName: "config") pod "94715e15-b083-4428-95d6-d7e66990af08" (UID: "94715e15-b083-4428-95d6-d7e66990af08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.282524 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-log-httpd\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.282588 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.282637 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-scripts\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.282699 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-run-httpd\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.282724 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.282756 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr26x\" (UniqueName: \"kubernetes.io/projected/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-kube-api-access-qr26x\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.282778 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-config-data\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.282851 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.282864 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpjjq\" (UniqueName: \"kubernetes.io/projected/94715e15-b083-4428-95d6-d7e66990af08-kube-api-access-gpjjq\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.282876 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.282884 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.284689 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-log-httpd\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.285233 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-run-httpd\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.287338 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-scripts\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.289284 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-config-data\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.290508 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.290713 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.298612 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "94715e15-b083-4428-95d6-d7e66990af08" (UID: "94715e15-b083-4428-95d6-d7e66990af08"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.300657 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr26x\" (UniqueName: \"kubernetes.io/projected/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-kube-api-access-qr26x\") pod \"ceilometer-0\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.385013 4744 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/94715e15-b083-4428-95d6-d7e66990af08-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.458423 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:30:10 crc kubenswrapper[4744]: I1008 09:30:10.928672 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:11 crc kubenswrapper[4744]: I1008 09:30:11.045662 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4txff" event={"ID":"93968d7d-7816-4580-8a04-f4ccacd584b6","Type":"ContainerStarted","Data":"b7479abcfb5a3ab44ee36628c757a9a869e08ea40da65001da7a1f10cd8b48ab"} Oct 08 09:30:11 crc kubenswrapper[4744]: I1008 09:30:11.048140 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f466fb8c4-5nw45" event={"ID":"94715e15-b083-4428-95d6-d7e66990af08","Type":"ContainerDied","Data":"0b6d1d92f1dbf328db183f181ed6c7c9c3dc0b4c146b9ac4dcd3e59e65441d68"} Oct 08 09:30:11 crc kubenswrapper[4744]: I1008 09:30:11.048453 4744 scope.go:117] "RemoveContainer" containerID="cd4ce2223ffbac03d43a8d5f76019aaa87e93027c60ead4fd2f9caaa285e42cb" Oct 08 09:30:11 crc kubenswrapper[4744]: I1008 09:30:11.048659 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f466fb8c4-5nw45" Oct 08 09:30:11 crc kubenswrapper[4744]: I1008 09:30:11.053222 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82","Type":"ContainerStarted","Data":"03122f65b6888322d086d124b7a25c601b35480513c8930fd0cf4737ff63fe9a"} Oct 08 09:30:11 crc kubenswrapper[4744]: I1008 09:30:11.079558 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-4txff" podStartSLOduration=2.601503832 podStartE2EDuration="17.079534517s" podCreationTimestamp="2025-10-08 09:29:54 +0000 UTC" firstStartedPulling="2025-10-08 09:29:55.767578448 +0000 UTC m=+1091.015223687" lastFinishedPulling="2025-10-08 09:30:10.245609133 +0000 UTC m=+1105.493254372" observedRunningTime="2025-10-08 09:30:11.073228978 +0000 UTC m=+1106.320874227" watchObservedRunningTime="2025-10-08 09:30:11.079534517 +0000 UTC m=+1106.327179746" Oct 08 09:30:11 crc kubenswrapper[4744]: I1008 09:30:11.094528 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7f466fb8c4-5nw45"] Oct 08 09:30:11 crc kubenswrapper[4744]: I1008 09:30:11.095651 4744 scope.go:117] "RemoveContainer" containerID="dbafe4df8d4eb5e4343d7baac824d5f64992310c6264c9583d609941e07ebfd7" Oct 08 09:30:11 crc kubenswrapper[4744]: I1008 09:30:11.098037 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7f466fb8c4-5nw45"] Oct 08 09:30:11 crc kubenswrapper[4744]: I1008 09:30:11.466570 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94715e15-b083-4428-95d6-d7e66990af08" path="/var/lib/kubelet/pods/94715e15-b083-4428-95d6-d7e66990af08/volumes" Oct 08 09:30:11 crc kubenswrapper[4744]: I1008 09:30:11.467360 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88fa333-9e1b-4ccb-a708-a256cd4f64e2" path="/var/lib/kubelet/pods/f88fa333-9e1b-4ccb-a708-a256cd4f64e2/volumes" Oct 08 09:30:12 crc kubenswrapper[4744]: I1008 09:30:12.069117 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82","Type":"ContainerStarted","Data":"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33"} Oct 08 09:30:14 crc kubenswrapper[4744]: I1008 09:30:14.112078 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82","Type":"ContainerStarted","Data":"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868"} Oct 08 09:30:15 crc kubenswrapper[4744]: I1008 09:30:15.125093 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82","Type":"ContainerStarted","Data":"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76"} Oct 08 09:30:16 crc kubenswrapper[4744]: I1008 09:30:16.136392 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82","Type":"ContainerStarted","Data":"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7"} Oct 08 09:30:16 crc kubenswrapper[4744]: I1008 09:30:16.138129 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 09:30:16 crc kubenswrapper[4744]: I1008 09:30:16.164294 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.864819998 podStartE2EDuration="6.164272224s" podCreationTimestamp="2025-10-08 09:30:10 +0000 UTC" firstStartedPulling="2025-10-08 09:30:10.938717342 +0000 UTC m=+1106.186362581" lastFinishedPulling="2025-10-08 09:30:15.238169568 +0000 UTC m=+1110.485814807" observedRunningTime="2025-10-08 09:30:16.160928705 +0000 UTC m=+1111.408573964" watchObservedRunningTime="2025-10-08 09:30:16.164272224 +0000 UTC m=+1111.411917463" Oct 08 09:30:19 crc kubenswrapper[4744]: I1008 09:30:19.690995 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:30:19 crc kubenswrapper[4744]: I1008 09:30:19.691349 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:30:20 crc kubenswrapper[4744]: I1008 09:30:20.343541 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:20 crc kubenswrapper[4744]: I1008 09:30:20.344431 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="ceilometer-central-agent" containerID="cri-o://2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33" gracePeriod=30 Oct 08 09:30:20 crc kubenswrapper[4744]: I1008 09:30:20.344530 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="ceilometer-notification-agent" containerID="cri-o://fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868" gracePeriod=30 Oct 08 09:30:20 crc kubenswrapper[4744]: I1008 09:30:20.344484 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="sg-core" containerID="cri-o://2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76" gracePeriod=30 Oct 08 09:30:20 crc kubenswrapper[4744]: I1008 09:30:20.344533 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="proxy-httpd" containerID="cri-o://b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7" gracePeriod=30 Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.085676 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.216529 4744 generic.go:334] "Generic (PLEG): container finished" podID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerID="b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7" exitCode=0 Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.216567 4744 generic.go:334] "Generic (PLEG): container finished" podID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerID="2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76" exitCode=2 Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.216578 4744 generic.go:334] "Generic (PLEG): container finished" podID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerID="fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868" exitCode=0 Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.216585 4744 generic.go:334] "Generic (PLEG): container finished" podID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerID="2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33" exitCode=0 Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.216605 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.216608 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82","Type":"ContainerDied","Data":"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7"} Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.216738 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82","Type":"ContainerDied","Data":"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76"} Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.216751 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82","Type":"ContainerDied","Data":"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868"} Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.216761 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82","Type":"ContainerDied","Data":"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33"} Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.216770 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82","Type":"ContainerDied","Data":"03122f65b6888322d086d124b7a25c601b35480513c8930fd0cf4737ff63fe9a"} Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.216786 4744 scope.go:117] "RemoveContainer" containerID="b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.224410 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-scripts\") pod \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.225765 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-sg-core-conf-yaml\") pod \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.226042 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-combined-ca-bundle\") pod \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.226135 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-log-httpd\") pod \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.226282 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-run-httpd\") pod \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.226440 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-config-data\") pod \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.226494 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr26x\" (UniqueName: \"kubernetes.io/projected/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-kube-api-access-qr26x\") pod \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\" (UID: \"4db36c6b-2d10-4b6e-b1e2-7af010f1ad82\") " Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.227013 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" (UID: "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.227932 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.228258 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" (UID: "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.238088 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-kube-api-access-qr26x" (OuterVolumeSpecName: "kube-api-access-qr26x") pod "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" (UID: "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82"). InnerVolumeSpecName "kube-api-access-qr26x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.240257 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-scripts" (OuterVolumeSpecName: "scripts") pod "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" (UID: "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.246416 4744 scope.go:117] "RemoveContainer" containerID="2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.263427 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" (UID: "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.265319 4744 scope.go:117] "RemoveContainer" containerID="fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.287914 4744 scope.go:117] "RemoveContainer" containerID="2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.308532 4744 scope.go:117] "RemoveContainer" containerID="b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7" Oct 08 09:30:21 crc kubenswrapper[4744]: E1008 09:30:21.309510 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7\": container with ID starting with b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7 not found: ID does not exist" containerID="b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.309545 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7"} err="failed to get container status \"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7\": rpc error: code = NotFound desc = could not find container \"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7\": container with ID starting with b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.309571 4744 scope.go:117] "RemoveContainer" containerID="2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76" Oct 08 09:30:21 crc kubenswrapper[4744]: E1008 09:30:21.310017 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76\": container with ID starting with 2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76 not found: ID does not exist" containerID="2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.310071 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76"} err="failed to get container status \"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76\": rpc error: code = NotFound desc = could not find container \"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76\": container with ID starting with 2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.310107 4744 scope.go:117] "RemoveContainer" containerID="fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868" Oct 08 09:30:21 crc kubenswrapper[4744]: E1008 09:30:21.310419 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868\": container with ID starting with fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868 not found: ID does not exist" containerID="fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.310439 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868"} err="failed to get container status \"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868\": rpc error: code = NotFound desc = could not find container \"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868\": container with ID starting with fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.310455 4744 scope.go:117] "RemoveContainer" containerID="2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33" Oct 08 09:30:21 crc kubenswrapper[4744]: E1008 09:30:21.310839 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33\": container with ID starting with 2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33 not found: ID does not exist" containerID="2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.310871 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33"} err="failed to get container status \"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33\": rpc error: code = NotFound desc = could not find container \"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33\": container with ID starting with 2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.310891 4744 scope.go:117] "RemoveContainer" containerID="b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.311136 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7"} err="failed to get container status \"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7\": rpc error: code = NotFound desc = could not find container \"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7\": container with ID starting with b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.311155 4744 scope.go:117] "RemoveContainer" containerID="2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.311436 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76"} err="failed to get container status \"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76\": rpc error: code = NotFound desc = could not find container \"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76\": container with ID starting with 2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.311459 4744 scope.go:117] "RemoveContainer" containerID="fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.311811 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868"} err="failed to get container status \"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868\": rpc error: code = NotFound desc = could not find container \"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868\": container with ID starting with fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.311861 4744 scope.go:117] "RemoveContainer" containerID="2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.312184 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33"} err="failed to get container status \"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33\": rpc error: code = NotFound desc = could not find container \"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33\": container with ID starting with 2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.312207 4744 scope.go:117] "RemoveContainer" containerID="b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.312960 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7"} err="failed to get container status \"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7\": rpc error: code = NotFound desc = could not find container \"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7\": container with ID starting with b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.312983 4744 scope.go:117] "RemoveContainer" containerID="2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.313261 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76"} err="failed to get container status \"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76\": rpc error: code = NotFound desc = could not find container \"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76\": container with ID starting with 2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.313311 4744 scope.go:117] "RemoveContainer" containerID="fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.313598 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868"} err="failed to get container status \"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868\": rpc error: code = NotFound desc = could not find container \"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868\": container with ID starting with fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.313620 4744 scope.go:117] "RemoveContainer" containerID="2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.313868 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33"} err="failed to get container status \"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33\": rpc error: code = NotFound desc = could not find container \"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33\": container with ID starting with 2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.313888 4744 scope.go:117] "RemoveContainer" containerID="b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.314251 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7"} err="failed to get container status \"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7\": rpc error: code = NotFound desc = could not find container \"b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7\": container with ID starting with b9f865936b002882e9c1614fd2e94969ef2f9a0c6e53cd0e3b7dddeffb02a1c7 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.314273 4744 scope.go:117] "RemoveContainer" containerID="2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.314660 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76"} err="failed to get container status \"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76\": rpc error: code = NotFound desc = could not find container \"2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76\": container with ID starting with 2f18cb5869773458d60a39961436cfc5359ba1e9d5bd6d2822c6f28caa263f76 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.314681 4744 scope.go:117] "RemoveContainer" containerID="fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.314992 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868"} err="failed to get container status \"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868\": rpc error: code = NotFound desc = could not find container \"fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868\": container with ID starting with fb30ad5076e00457d8e2454a78d608f0e7a209887f85fde8b34b0f48d50ef868 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.315018 4744 scope.go:117] "RemoveContainer" containerID="2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.315543 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33"} err="failed to get container status \"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33\": rpc error: code = NotFound desc = could not find container \"2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33\": container with ID starting with 2d0f850939985054653bf9dcfee531abcf0eb6e3060c5cdf69efa0287e4f1e33 not found: ID does not exist" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.320420 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" (UID: "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.329730 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.329756 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.329767 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr26x\" (UniqueName: \"kubernetes.io/projected/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-kube-api-access-qr26x\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.329780 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.329789 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.340320 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-config-data" (OuterVolumeSpecName: "config-data") pod "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" (UID: "4db36c6b-2d10-4b6e-b1e2-7af010f1ad82"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.431260 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.541895 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.547562 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.576502 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:21 crc kubenswrapper[4744]: E1008 09:30:21.576988 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="ceilometer-notification-agent" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.577014 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="ceilometer-notification-agent" Oct 08 09:30:21 crc kubenswrapper[4744]: E1008 09:30:21.577053 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="ceilometer-central-agent" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.577062 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="ceilometer-central-agent" Oct 08 09:30:21 crc kubenswrapper[4744]: E1008 09:30:21.577077 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="proxy-httpd" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.577085 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="proxy-httpd" Oct 08 09:30:21 crc kubenswrapper[4744]: E1008 09:30:21.577104 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="sg-core" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.577115 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="sg-core" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.577351 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="sg-core" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.577401 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="ceilometer-central-agent" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.577429 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="ceilometer-notification-agent" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.577447 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" containerName="proxy-httpd" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.583228 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.585558 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.585830 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.602058 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.634027 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-config-data\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.634194 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pjqw\" (UniqueName: \"kubernetes.io/projected/68b2d9e7-2617-4c74-b4cd-ee619204ee07-kube-api-access-9pjqw\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.634227 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b2d9e7-2617-4c74-b4cd-ee619204ee07-run-httpd\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.634394 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-scripts\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.634431 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.634456 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b2d9e7-2617-4c74-b4cd-ee619204ee07-log-httpd\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.634480 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.736487 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-config-data\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.737129 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pjqw\" (UniqueName: \"kubernetes.io/projected/68b2d9e7-2617-4c74-b4cd-ee619204ee07-kube-api-access-9pjqw\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.737164 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b2d9e7-2617-4c74-b4cd-ee619204ee07-run-httpd\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.737598 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b2d9e7-2617-4c74-b4cd-ee619204ee07-run-httpd\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.737828 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-scripts\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.737857 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.738205 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b2d9e7-2617-4c74-b4cd-ee619204ee07-log-httpd\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.738236 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.738706 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b2d9e7-2617-4c74-b4cd-ee619204ee07-log-httpd\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.742068 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-config-data\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.742617 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-scripts\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.744727 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.750840 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.757180 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pjqw\" (UniqueName: \"kubernetes.io/projected/68b2d9e7-2617-4c74-b4cd-ee619204ee07-kube-api-access-9pjqw\") pod \"ceilometer-0\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " pod="openstack/ceilometer-0" Oct 08 09:30:21 crc kubenswrapper[4744]: I1008 09:30:21.901872 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:30:22 crc kubenswrapper[4744]: I1008 09:30:22.394761 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:22 crc kubenswrapper[4744]: W1008 09:30:22.405816 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68b2d9e7_2617_4c74_b4cd_ee619204ee07.slice/crio-7f6979499ef75886cf5ba2776c757ad96ce2a025e88276cc6dd1555beea6f433 WatchSource:0}: Error finding container 7f6979499ef75886cf5ba2776c757ad96ce2a025e88276cc6dd1555beea6f433: Status 404 returned error can't find the container with id 7f6979499ef75886cf5ba2776c757ad96ce2a025e88276cc6dd1555beea6f433 Oct 08 09:30:23 crc kubenswrapper[4744]: I1008 09:30:23.238110 4744 generic.go:334] "Generic (PLEG): container finished" podID="93968d7d-7816-4580-8a04-f4ccacd584b6" containerID="b7479abcfb5a3ab44ee36628c757a9a869e08ea40da65001da7a1f10cd8b48ab" exitCode=0 Oct 08 09:30:23 crc kubenswrapper[4744]: I1008 09:30:23.238527 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4txff" event={"ID":"93968d7d-7816-4580-8a04-f4ccacd584b6","Type":"ContainerDied","Data":"b7479abcfb5a3ab44ee36628c757a9a869e08ea40da65001da7a1f10cd8b48ab"} Oct 08 09:30:23 crc kubenswrapper[4744]: I1008 09:30:23.246134 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b2d9e7-2617-4c74-b4cd-ee619204ee07","Type":"ContainerStarted","Data":"bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28"} Oct 08 09:30:23 crc kubenswrapper[4744]: I1008 09:30:23.246209 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b2d9e7-2617-4c74-b4cd-ee619204ee07","Type":"ContainerStarted","Data":"7f6979499ef75886cf5ba2776c757ad96ce2a025e88276cc6dd1555beea6f433"} Oct 08 09:30:23 crc kubenswrapper[4744]: I1008 09:30:23.462883 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4db36c6b-2d10-4b6e-b1e2-7af010f1ad82" path="/var/lib/kubelet/pods/4db36c6b-2d10-4b6e-b1e2-7af010f1ad82/volumes" Oct 08 09:30:24 crc kubenswrapper[4744]: I1008 09:30:24.259251 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b2d9e7-2617-4c74-b4cd-ee619204ee07","Type":"ContainerStarted","Data":"217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92"} Oct 08 09:30:24 crc kubenswrapper[4744]: I1008 09:30:24.624493 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:30:24 crc kubenswrapper[4744]: I1008 09:30:24.695080 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-scripts\") pod \"93968d7d-7816-4580-8a04-f4ccacd584b6\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " Oct 08 09:30:24 crc kubenswrapper[4744]: I1008 09:30:24.695161 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt254\" (UniqueName: \"kubernetes.io/projected/93968d7d-7816-4580-8a04-f4ccacd584b6-kube-api-access-tt254\") pod \"93968d7d-7816-4580-8a04-f4ccacd584b6\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " Oct 08 09:30:24 crc kubenswrapper[4744]: I1008 09:30:24.695201 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-combined-ca-bundle\") pod \"93968d7d-7816-4580-8a04-f4ccacd584b6\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " Oct 08 09:30:24 crc kubenswrapper[4744]: I1008 09:30:24.695300 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-config-data\") pod \"93968d7d-7816-4580-8a04-f4ccacd584b6\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " Oct 08 09:30:24 crc kubenswrapper[4744]: I1008 09:30:24.705514 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-scripts" (OuterVolumeSpecName: "scripts") pod "93968d7d-7816-4580-8a04-f4ccacd584b6" (UID: "93968d7d-7816-4580-8a04-f4ccacd584b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:24 crc kubenswrapper[4744]: I1008 09:30:24.705723 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93968d7d-7816-4580-8a04-f4ccacd584b6-kube-api-access-tt254" (OuterVolumeSpecName: "kube-api-access-tt254") pod "93968d7d-7816-4580-8a04-f4ccacd584b6" (UID: "93968d7d-7816-4580-8a04-f4ccacd584b6"). InnerVolumeSpecName "kube-api-access-tt254". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:24 crc kubenswrapper[4744]: E1008 09:30:24.723164 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-config-data podName:93968d7d-7816-4580-8a04-f4ccacd584b6 nodeName:}" failed. No retries permitted until 2025-10-08 09:30:25.223131364 +0000 UTC m=+1120.470776603 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-config-data") pod "93968d7d-7816-4580-8a04-f4ccacd584b6" (UID: "93968d7d-7816-4580-8a04-f4ccacd584b6") : error deleting /var/lib/kubelet/pods/93968d7d-7816-4580-8a04-f4ccacd584b6/volume-subpaths: remove /var/lib/kubelet/pods/93968d7d-7816-4580-8a04-f4ccacd584b6/volume-subpaths: no such file or directory Oct 08 09:30:24 crc kubenswrapper[4744]: I1008 09:30:24.725747 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93968d7d-7816-4580-8a04-f4ccacd584b6" (UID: "93968d7d-7816-4580-8a04-f4ccacd584b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:24 crc kubenswrapper[4744]: I1008 09:30:24.797180 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt254\" (UniqueName: \"kubernetes.io/projected/93968d7d-7816-4580-8a04-f4ccacd584b6-kube-api-access-tt254\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:24 crc kubenswrapper[4744]: I1008 09:30:24.797213 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:24 crc kubenswrapper[4744]: I1008 09:30:24.797222 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.270040 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-4txff" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.270079 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-4txff" event={"ID":"93968d7d-7816-4580-8a04-f4ccacd584b6","Type":"ContainerDied","Data":"7eb8561cd2f6878e0ed559766b692a2af187855fd1c06e31b358d7492d933022"} Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.270556 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7eb8561cd2f6878e0ed559766b692a2af187855fd1c06e31b358d7492d933022" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.273784 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b2d9e7-2617-4c74-b4cd-ee619204ee07","Type":"ContainerStarted","Data":"f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33"} Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.306931 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-config-data\") pod \"93968d7d-7816-4580-8a04-f4ccacd584b6\" (UID: \"93968d7d-7816-4580-8a04-f4ccacd584b6\") " Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.316625 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-config-data" (OuterVolumeSpecName: "config-data") pod "93968d7d-7816-4580-8a04-f4ccacd584b6" (UID: "93968d7d-7816-4580-8a04-f4ccacd584b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.354590 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 09:30:25 crc kubenswrapper[4744]: E1008 09:30:25.355148 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93968d7d-7816-4580-8a04-f4ccacd584b6" containerName="nova-cell0-conductor-db-sync" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.355168 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="93968d7d-7816-4580-8a04-f4ccacd584b6" containerName="nova-cell0-conductor-db-sync" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.355440 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="93968d7d-7816-4580-8a04-f4ccacd584b6" containerName="nova-cell0-conductor-db-sync" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.356239 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.381397 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.409441 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p97nf\" (UniqueName: \"kubernetes.io/projected/3e036c58-fa76-41bb-a16f-77566f7e5533-kube-api-access-p97nf\") pod \"nova-cell0-conductor-0\" (UID: \"3e036c58-fa76-41bb-a16f-77566f7e5533\") " pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.409519 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e036c58-fa76-41bb-a16f-77566f7e5533-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3e036c58-fa76-41bb-a16f-77566f7e5533\") " pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.409690 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e036c58-fa76-41bb-a16f-77566f7e5533-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3e036c58-fa76-41bb-a16f-77566f7e5533\") " pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.409782 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93968d7d-7816-4580-8a04-f4ccacd584b6-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.511555 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p97nf\" (UniqueName: \"kubernetes.io/projected/3e036c58-fa76-41bb-a16f-77566f7e5533-kube-api-access-p97nf\") pod \"nova-cell0-conductor-0\" (UID: \"3e036c58-fa76-41bb-a16f-77566f7e5533\") " pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.511662 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e036c58-fa76-41bb-a16f-77566f7e5533-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3e036c58-fa76-41bb-a16f-77566f7e5533\") " pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.512685 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e036c58-fa76-41bb-a16f-77566f7e5533-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3e036c58-fa76-41bb-a16f-77566f7e5533\") " pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.517893 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e036c58-fa76-41bb-a16f-77566f7e5533-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3e036c58-fa76-41bb-a16f-77566f7e5533\") " pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.518546 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e036c58-fa76-41bb-a16f-77566f7e5533-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3e036c58-fa76-41bb-a16f-77566f7e5533\") " pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.544067 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p97nf\" (UniqueName: \"kubernetes.io/projected/3e036c58-fa76-41bb-a16f-77566f7e5533-kube-api-access-p97nf\") pod \"nova-cell0-conductor-0\" (UID: \"3e036c58-fa76-41bb-a16f-77566f7e5533\") " pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:25 crc kubenswrapper[4744]: I1008 09:30:25.693065 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:26 crc kubenswrapper[4744]: I1008 09:30:26.248510 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 09:30:26 crc kubenswrapper[4744]: I1008 09:30:26.296152 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b2d9e7-2617-4c74-b4cd-ee619204ee07","Type":"ContainerStarted","Data":"bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6"} Oct 08 09:30:26 crc kubenswrapper[4744]: I1008 09:30:26.297405 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 09:30:26 crc kubenswrapper[4744]: I1008 09:30:26.304904 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3e036c58-fa76-41bb-a16f-77566f7e5533","Type":"ContainerStarted","Data":"8832ff3c8e43a79ef92d733b51105647615f534ae3b75c409c2cffd9b8ab597c"} Oct 08 09:30:26 crc kubenswrapper[4744]: I1008 09:30:26.326564 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.065173803 podStartE2EDuration="5.326538608s" podCreationTimestamp="2025-10-08 09:30:21 +0000 UTC" firstStartedPulling="2025-10-08 09:30:22.414807266 +0000 UTC m=+1117.662452506" lastFinishedPulling="2025-10-08 09:30:25.676172072 +0000 UTC m=+1120.923817311" observedRunningTime="2025-10-08 09:30:26.324168554 +0000 UTC m=+1121.571813793" watchObservedRunningTime="2025-10-08 09:30:26.326538608 +0000 UTC m=+1121.574183847" Oct 08 09:30:27 crc kubenswrapper[4744]: I1008 09:30:27.316521 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3e036c58-fa76-41bb-a16f-77566f7e5533","Type":"ContainerStarted","Data":"541c3dc6d3d38bcd25c5f8e621b29f213f97ff2577612fc2ffa86a3f15b180a5"} Oct 08 09:30:27 crc kubenswrapper[4744]: I1008 09:30:27.316959 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:27 crc kubenswrapper[4744]: I1008 09:30:27.340962 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.340941855 podStartE2EDuration="2.340941855s" podCreationTimestamp="2025-10-08 09:30:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:30:27.335012227 +0000 UTC m=+1122.582657466" watchObservedRunningTime="2025-10-08 09:30:27.340941855 +0000 UTC m=+1122.588587094" Oct 08 09:30:35 crc kubenswrapper[4744]: I1008 09:30:35.724826 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.257779 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-x4khv"] Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.274192 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.276806 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.281025 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.292080 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x4khv"] Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.349608 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-scripts\") pod \"nova-cell0-cell-mapping-x4khv\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.349663 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-config-data\") pod \"nova-cell0-cell-mapping-x4khv\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.349739 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jlm9\" (UniqueName: \"kubernetes.io/projected/f8711d76-061a-4b5a-a2f6-05996808b79d-kube-api-access-2jlm9\") pod \"nova-cell0-cell-mapping-x4khv\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.349764 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x4khv\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.457478 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.483453 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-scripts\") pod \"nova-cell0-cell-mapping-x4khv\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.490517 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-config-data\") pod \"nova-cell0-cell-mapping-x4khv\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.491105 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jlm9\" (UniqueName: \"kubernetes.io/projected/f8711d76-061a-4b5a-a2f6-05996808b79d-kube-api-access-2jlm9\") pod \"nova-cell0-cell-mapping-x4khv\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.491386 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x4khv\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.505859 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.545341 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.558750 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-scripts\") pod \"nova-cell0-cell-mapping-x4khv\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.607290 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x4khv\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.608062 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-config-data\") pod \"nova-cell0-cell-mapping-x4khv\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.614323 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.614985 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqkvs\" (UniqueName: \"kubernetes.io/projected/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-kube-api-access-nqkvs\") pod \"nova-scheduler-0\" (UID: \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.615053 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-config-data\") pod \"nova-scheduler-0\" (UID: \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.636434 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.639065 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jlm9\" (UniqueName: \"kubernetes.io/projected/f8711d76-061a-4b5a-a2f6-05996808b79d-kube-api-access-2jlm9\") pod \"nova-cell0-cell-mapping-x4khv\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.652960 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.654915 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.669635 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.676424 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.717269 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab3407e4-f332-4de8-953a-e92ed39c4e7b-logs\") pod \"nova-api-0\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.721265 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.721530 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3407e4-f332-4de8-953a-e92ed39c4e7b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.721812 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfkds\" (UniqueName: \"kubernetes.io/projected/ab3407e4-f332-4de8-953a-e92ed39c4e7b-kube-api-access-bfkds\") pod \"nova-api-0\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.721901 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3407e4-f332-4de8-953a-e92ed39c4e7b-config-data\") pod \"nova-api-0\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.721977 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqkvs\" (UniqueName: \"kubernetes.io/projected/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-kube-api-access-nqkvs\") pod \"nova-scheduler-0\" (UID: \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.722043 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-config-data\") pod \"nova-scheduler-0\" (UID: \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.732257 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.733077 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-config-data\") pod \"nova-scheduler-0\" (UID: \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.733712 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.740297 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.753905 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.756456 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.801517 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqkvs\" (UniqueName: \"kubernetes.io/projected/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-kube-api-access-nqkvs\") pod \"nova-scheduler-0\" (UID: \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.823789 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e7a440-aafd-4452-9e48-bc4bf74980ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " pod="openstack/nova-metadata-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.823835 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hwxt\" (UniqueName: \"kubernetes.io/projected/e0e7a440-aafd-4452-9e48-bc4bf74980ab-kube-api-access-2hwxt\") pod \"nova-metadata-0\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " pod="openstack/nova-metadata-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.823920 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3407e4-f332-4de8-953a-e92ed39c4e7b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.823946 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e7a440-aafd-4452-9e48-bc4bf74980ab-config-data\") pod \"nova-metadata-0\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " pod="openstack/nova-metadata-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.823985 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfkds\" (UniqueName: \"kubernetes.io/projected/ab3407e4-f332-4de8-953a-e92ed39c4e7b-kube-api-access-bfkds\") pod \"nova-api-0\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.824009 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0e7a440-aafd-4452-9e48-bc4bf74980ab-logs\") pod \"nova-metadata-0\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " pod="openstack/nova-metadata-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.824025 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3407e4-f332-4de8-953a-e92ed39c4e7b-config-data\") pod \"nova-api-0\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.824080 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab3407e4-f332-4de8-953a-e92ed39c4e7b-logs\") pod \"nova-api-0\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.825447 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab3407e4-f332-4de8-953a-e92ed39c4e7b-logs\") pod \"nova-api-0\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.834087 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3407e4-f332-4de8-953a-e92ed39c4e7b-config-data\") pod \"nova-api-0\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.837184 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3407e4-f332-4de8-953a-e92ed39c4e7b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.852030 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75fb48c489-dbtdd"] Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.856396 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.871333 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.876044 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfkds\" (UniqueName: \"kubernetes.io/projected/ab3407e4-f332-4de8-953a-e92ed39c4e7b-kube-api-access-bfkds\") pod \"nova-api-0\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " pod="openstack/nova-api-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.877804 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.892667 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.901727 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.905051 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75fb48c489-dbtdd"] Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.925725 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6156b6be-400b-45e6-afee-7fc443b2b23d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6156b6be-400b-45e6-afee-7fc443b2b23d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.925819 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e7a440-aafd-4452-9e48-bc4bf74980ab-config-data\") pod \"nova-metadata-0\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " pod="openstack/nova-metadata-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.925843 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6156b6be-400b-45e6-afee-7fc443b2b23d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6156b6be-400b-45e6-afee-7fc443b2b23d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.925861 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-ovsdbserver-nb\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.925906 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-config\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.925937 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0e7a440-aafd-4452-9e48-bc4bf74980ab-logs\") pod \"nova-metadata-0\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " pod="openstack/nova-metadata-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.925962 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcjkg\" (UniqueName: \"kubernetes.io/projected/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-kube-api-access-dcjkg\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.926013 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9bwx\" (UniqueName: \"kubernetes.io/projected/6156b6be-400b-45e6-afee-7fc443b2b23d-kube-api-access-h9bwx\") pod \"nova-cell1-novncproxy-0\" (UID: \"6156b6be-400b-45e6-afee-7fc443b2b23d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.926050 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e7a440-aafd-4452-9e48-bc4bf74980ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " pod="openstack/nova-metadata-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.926068 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hwxt\" (UniqueName: \"kubernetes.io/projected/e0e7a440-aafd-4452-9e48-bc4bf74980ab-kube-api-access-2hwxt\") pod \"nova-metadata-0\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " pod="openstack/nova-metadata-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.926082 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-ovsdbserver-sb\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.926119 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-dns-svc\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.934194 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e7a440-aafd-4452-9e48-bc4bf74980ab-config-data\") pod \"nova-metadata-0\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " pod="openstack/nova-metadata-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.934525 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0e7a440-aafd-4452-9e48-bc4bf74980ab-logs\") pod \"nova-metadata-0\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " pod="openstack/nova-metadata-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.940298 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e7a440-aafd-4452-9e48-bc4bf74980ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " pod="openstack/nova-metadata-0" Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.957483 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 09:30:36 crc kubenswrapper[4744]: I1008 09:30:36.974542 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hwxt\" (UniqueName: \"kubernetes.io/projected/e0e7a440-aafd-4452-9e48-bc4bf74980ab-kube-api-access-2hwxt\") pod \"nova-metadata-0\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " pod="openstack/nova-metadata-0" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.028826 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.029457 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6156b6be-400b-45e6-afee-7fc443b2b23d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6156b6be-400b-45e6-afee-7fc443b2b23d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.029544 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6156b6be-400b-45e6-afee-7fc443b2b23d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6156b6be-400b-45e6-afee-7fc443b2b23d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.029577 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-ovsdbserver-nb\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.030151 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-config\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.030202 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcjkg\" (UniqueName: \"kubernetes.io/projected/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-kube-api-access-dcjkg\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.030253 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9bwx\" (UniqueName: \"kubernetes.io/projected/6156b6be-400b-45e6-afee-7fc443b2b23d-kube-api-access-h9bwx\") pod \"nova-cell1-novncproxy-0\" (UID: \"6156b6be-400b-45e6-afee-7fc443b2b23d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.030299 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-ovsdbserver-sb\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.030342 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-dns-svc\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.030925 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-ovsdbserver-nb\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.030960 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-config\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.031547 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-dns-svc\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.031872 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-ovsdbserver-sb\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.034131 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6156b6be-400b-45e6-afee-7fc443b2b23d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6156b6be-400b-45e6-afee-7fc443b2b23d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.037999 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6156b6be-400b-45e6-afee-7fc443b2b23d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6156b6be-400b-45e6-afee-7fc443b2b23d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.054133 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9bwx\" (UniqueName: \"kubernetes.io/projected/6156b6be-400b-45e6-afee-7fc443b2b23d-kube-api-access-h9bwx\") pod \"nova-cell1-novncproxy-0\" (UID: \"6156b6be-400b-45e6-afee-7fc443b2b23d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.055528 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcjkg\" (UniqueName: \"kubernetes.io/projected/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-kube-api-access-dcjkg\") pod \"dnsmasq-dns-75fb48c489-dbtdd\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.086175 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.150768 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.186864 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.206862 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.533180 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x4khv"] Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.773814 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:30:37 crc kubenswrapper[4744]: W1008 09:30:37.950610 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab3407e4_f332_4de8_953a_e92ed39c4e7b.slice/crio-c14d146e6f86f7e8c76477c5941c5a6390f1ceb9010acc9239adbfe508ecbb54 WatchSource:0}: Error finding container c14d146e6f86f7e8c76477c5941c5a6390f1ceb9010acc9239adbfe508ecbb54: Status 404 returned error can't find the container with id c14d146e6f86f7e8c76477c5941c5a6390f1ceb9010acc9239adbfe508ecbb54 Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.954517 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:30:37 crc kubenswrapper[4744]: I1008 09:30:37.978481 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.002129 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75fb48c489-dbtdd"] Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.014859 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:38 crc kubenswrapper[4744]: W1008 09:30:38.015915 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0e7a440_aafd_4452_9e48_bc4bf74980ab.slice/crio-0bf17f18022ea73cc5abeb1b1866160b33dcb8c9d1384c6fc3e027f65b44eaac WatchSource:0}: Error finding container 0bf17f18022ea73cc5abeb1b1866160b33dcb8c9d1384c6fc3e027f65b44eaac: Status 404 returned error can't find the container with id 0bf17f18022ea73cc5abeb1b1866160b33dcb8c9d1384c6fc3e027f65b44eaac Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.032212 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vgxfh"] Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.033637 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.037287 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.037287 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.094599 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vgxfh"] Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.179067 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vgxfh\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.179125 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-config-data\") pod \"nova-cell1-conductor-db-sync-vgxfh\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.179162 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-scripts\") pod \"nova-cell1-conductor-db-sync-vgxfh\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.179194 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm89g\" (UniqueName: \"kubernetes.io/projected/6bc195e6-d464-4837-b6a0-f11b629e3446-kube-api-access-nm89g\") pod \"nova-cell1-conductor-db-sync-vgxfh\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.281748 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vgxfh\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.282062 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-config-data\") pod \"nova-cell1-conductor-db-sync-vgxfh\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.282165 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-scripts\") pod \"nova-cell1-conductor-db-sync-vgxfh\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.282275 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm89g\" (UniqueName: \"kubernetes.io/projected/6bc195e6-d464-4837-b6a0-f11b629e3446-kube-api-access-nm89g\") pod \"nova-cell1-conductor-db-sync-vgxfh\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.287270 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-config-data\") pod \"nova-cell1-conductor-db-sync-vgxfh\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.287707 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-scripts\") pod \"nova-cell1-conductor-db-sync-vgxfh\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.288322 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vgxfh\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.298953 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm89g\" (UniqueName: \"kubernetes.io/projected/6bc195e6-d464-4837-b6a0-f11b629e3446-kube-api-access-nm89g\") pod \"nova-cell1-conductor-db-sync-vgxfh\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.384189 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.547460 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x4khv" event={"ID":"f8711d76-061a-4b5a-a2f6-05996808b79d","Type":"ContainerStarted","Data":"2f8b94fa09ec47e8aeb6aab803b6abe46eea297c395a4af7d144259c22f62236"} Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.549050 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x4khv" event={"ID":"f8711d76-061a-4b5a-a2f6-05996808b79d","Type":"ContainerStarted","Data":"57cf58d19c2ced2c9ee8e42546e3a7c6a87928d26ad4aeaf0087958768d5e796"} Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.553793 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ab3407e4-f332-4de8-953a-e92ed39c4e7b","Type":"ContainerStarted","Data":"c14d146e6f86f7e8c76477c5941c5a6390f1ceb9010acc9239adbfe508ecbb54"} Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.555614 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6156b6be-400b-45e6-afee-7fc443b2b23d","Type":"ContainerStarted","Data":"89da7186eb9d228e6a6ec6eea18cf216a27b403d113fe8f6252f5700ae97b935"} Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.580726 4744 generic.go:334] "Generic (PLEG): container finished" podID="a81b3a01-b0c5-4a5f-b50e-6c28837ff977" containerID="f76d20950f06678527ff0d329d8b1b63bd12a28221fe466d63e6c07c89bc3370" exitCode=0 Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.580934 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" event={"ID":"a81b3a01-b0c5-4a5f-b50e-6c28837ff977","Type":"ContainerDied","Data":"f76d20950f06678527ff0d329d8b1b63bd12a28221fe466d63e6c07c89bc3370"} Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.581001 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" event={"ID":"a81b3a01-b0c5-4a5f-b50e-6c28837ff977","Type":"ContainerStarted","Data":"024b2df5a5d0e337de61d42f8330eb105bff527548ae5deab74b3d1a1c898b4e"} Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.588775 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-x4khv" podStartSLOduration=2.588744037 podStartE2EDuration="2.588744037s" podCreationTimestamp="2025-10-08 09:30:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:30:38.563755158 +0000 UTC m=+1133.811400397" watchObservedRunningTime="2025-10-08 09:30:38.588744037 +0000 UTC m=+1133.836389276" Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.605111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0e7a440-aafd-4452-9e48-bc4bf74980ab","Type":"ContainerStarted","Data":"0bf17f18022ea73cc5abeb1b1866160b33dcb8c9d1384c6fc3e027f65b44eaac"} Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.606082 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a","Type":"ContainerStarted","Data":"82154c8240c2c8028252194c92e3ca26098a76d05736d01cf9e67016932e4f20"} Oct 08 09:30:38 crc kubenswrapper[4744]: I1008 09:30:38.744967 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vgxfh"] Oct 08 09:30:38 crc kubenswrapper[4744]: W1008 09:30:38.767771 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bc195e6_d464_4837_b6a0_f11b629e3446.slice/crio-e8da1283e79e847b1355c6e6ae5f83ddb3e34204776699b105b323166fabea8f WatchSource:0}: Error finding container e8da1283e79e847b1355c6e6ae5f83ddb3e34204776699b105b323166fabea8f: Status 404 returned error can't find the container with id e8da1283e79e847b1355c6e6ae5f83ddb3e34204776699b105b323166fabea8f Oct 08 09:30:39 crc kubenswrapper[4744]: I1008 09:30:39.651961 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" event={"ID":"a81b3a01-b0c5-4a5f-b50e-6c28837ff977","Type":"ContainerStarted","Data":"8f697f3d0acca97b854bba900c82fab6e7dee4be65cc9e848d46bf9382b30c00"} Oct 08 09:30:39 crc kubenswrapper[4744]: I1008 09:30:39.652071 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:39 crc kubenswrapper[4744]: I1008 09:30:39.666642 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vgxfh" event={"ID":"6bc195e6-d464-4837-b6a0-f11b629e3446","Type":"ContainerStarted","Data":"398904171cafba1a279cf46661a2150313ab50fe0b08ce7ff92fbab6592beb76"} Oct 08 09:30:39 crc kubenswrapper[4744]: I1008 09:30:39.666696 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vgxfh" event={"ID":"6bc195e6-d464-4837-b6a0-f11b629e3446","Type":"ContainerStarted","Data":"e8da1283e79e847b1355c6e6ae5f83ddb3e34204776699b105b323166fabea8f"} Oct 08 09:30:39 crc kubenswrapper[4744]: I1008 09:30:39.696962 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" podStartSLOduration=3.696941822 podStartE2EDuration="3.696941822s" podCreationTimestamp="2025-10-08 09:30:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:30:39.692698058 +0000 UTC m=+1134.940343307" watchObservedRunningTime="2025-10-08 09:30:39.696941822 +0000 UTC m=+1134.944587061" Oct 08 09:30:39 crc kubenswrapper[4744]: I1008 09:30:39.733632 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-vgxfh" podStartSLOduration=2.733605742 podStartE2EDuration="2.733605742s" podCreationTimestamp="2025-10-08 09:30:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:30:39.719766952 +0000 UTC m=+1134.967412201" watchObservedRunningTime="2025-10-08 09:30:39.733605742 +0000 UTC m=+1134.981250971" Oct 08 09:30:40 crc kubenswrapper[4744]: I1008 09:30:40.628913 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 09:30:40 crc kubenswrapper[4744]: I1008 09:30:40.665590 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.707047 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ab3407e4-f332-4de8-953a-e92ed39c4e7b","Type":"ContainerStarted","Data":"7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc"} Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.707448 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ab3407e4-f332-4de8-953a-e92ed39c4e7b","Type":"ContainerStarted","Data":"d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85"} Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.710487 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6156b6be-400b-45e6-afee-7fc443b2b23d","Type":"ContainerStarted","Data":"d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c"} Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.710664 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="6156b6be-400b-45e6-afee-7fc443b2b23d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c" gracePeriod=30 Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.716076 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0e7a440-aafd-4452-9e48-bc4bf74980ab","Type":"ContainerStarted","Data":"652899e54dcc6b2ad13dd42267f9a8e10cb6f8f9f0b0d585511ea9005450af3c"} Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.716600 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0e7a440-aafd-4452-9e48-bc4bf74980ab","Type":"ContainerStarted","Data":"cc71f0c8c72c15736d655cfbd3aae35933ef71569a56dff8053d12bb6374e1fa"} Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.716767 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e0e7a440-aafd-4452-9e48-bc4bf74980ab" containerName="nova-metadata-log" containerID="cri-o://cc71f0c8c72c15736d655cfbd3aae35933ef71569a56dff8053d12bb6374e1fa" gracePeriod=30 Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.716983 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e0e7a440-aafd-4452-9e48-bc4bf74980ab" containerName="nova-metadata-metadata" containerID="cri-o://652899e54dcc6b2ad13dd42267f9a8e10cb6f8f9f0b0d585511ea9005450af3c" gracePeriod=30 Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.721201 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a","Type":"ContainerStarted","Data":"13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594"} Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.748303 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.744911902 podStartE2EDuration="6.748279413s" podCreationTimestamp="2025-10-08 09:30:36 +0000 UTC" firstStartedPulling="2025-10-08 09:30:37.960099092 +0000 UTC m=+1133.207744331" lastFinishedPulling="2025-10-08 09:30:41.963466603 +0000 UTC m=+1137.211111842" observedRunningTime="2025-10-08 09:30:42.740594537 +0000 UTC m=+1137.988239776" watchObservedRunningTime="2025-10-08 09:30:42.748279413 +0000 UTC m=+1137.995924652" Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.758710 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.570460708 podStartE2EDuration="6.758689031s" podCreationTimestamp="2025-10-08 09:30:36 +0000 UTC" firstStartedPulling="2025-10-08 09:30:37.772867686 +0000 UTC m=+1133.020512925" lastFinishedPulling="2025-10-08 09:30:41.961096009 +0000 UTC m=+1137.208741248" observedRunningTime="2025-10-08 09:30:42.756991926 +0000 UTC m=+1138.004637175" watchObservedRunningTime="2025-10-08 09:30:42.758689031 +0000 UTC m=+1138.006334270" Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.785830 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.800627601 podStartE2EDuration="6.785809996s" podCreationTimestamp="2025-10-08 09:30:36 +0000 UTC" firstStartedPulling="2025-10-08 09:30:37.978141864 +0000 UTC m=+1133.225787103" lastFinishedPulling="2025-10-08 09:30:41.963324259 +0000 UTC m=+1137.210969498" observedRunningTime="2025-10-08 09:30:42.781771858 +0000 UTC m=+1138.029417097" watchObservedRunningTime="2025-10-08 09:30:42.785809996 +0000 UTC m=+1138.033455225" Oct 08 09:30:42 crc kubenswrapper[4744]: I1008 09:30:42.802174 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.831151527 podStartE2EDuration="6.802151493s" podCreationTimestamp="2025-10-08 09:30:36 +0000 UTC" firstStartedPulling="2025-10-08 09:30:38.023622879 +0000 UTC m=+1133.271268118" lastFinishedPulling="2025-10-08 09:30:41.994622845 +0000 UTC m=+1137.242268084" observedRunningTime="2025-10-08 09:30:42.800314963 +0000 UTC m=+1138.047960202" watchObservedRunningTime="2025-10-08 09:30:42.802151493 +0000 UTC m=+1138.049796722" Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.732600 4744 generic.go:334] "Generic (PLEG): container finished" podID="e0e7a440-aafd-4452-9e48-bc4bf74980ab" containerID="652899e54dcc6b2ad13dd42267f9a8e10cb6f8f9f0b0d585511ea9005450af3c" exitCode=0 Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.734049 4744 generic.go:334] "Generic (PLEG): container finished" podID="e0e7a440-aafd-4452-9e48-bc4bf74980ab" containerID="cc71f0c8c72c15736d655cfbd3aae35933ef71569a56dff8053d12bb6374e1fa" exitCode=143 Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.732695 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0e7a440-aafd-4452-9e48-bc4bf74980ab","Type":"ContainerDied","Data":"652899e54dcc6b2ad13dd42267f9a8e10cb6f8f9f0b0d585511ea9005450af3c"} Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.734191 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0e7a440-aafd-4452-9e48-bc4bf74980ab","Type":"ContainerDied","Data":"cc71f0c8c72c15736d655cfbd3aae35933ef71569a56dff8053d12bb6374e1fa"} Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.734213 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e0e7a440-aafd-4452-9e48-bc4bf74980ab","Type":"ContainerDied","Data":"0bf17f18022ea73cc5abeb1b1866160b33dcb8c9d1384c6fc3e027f65b44eaac"} Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.734223 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bf17f18022ea73cc5abeb1b1866160b33dcb8c9d1384c6fc3e027f65b44eaac" Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.753666 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.846945 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e7a440-aafd-4452-9e48-bc4bf74980ab-config-data\") pod \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.847300 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hwxt\" (UniqueName: \"kubernetes.io/projected/e0e7a440-aafd-4452-9e48-bc4bf74980ab-kube-api-access-2hwxt\") pod \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.847517 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0e7a440-aafd-4452-9e48-bc4bf74980ab-logs\") pod \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.847652 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e7a440-aafd-4452-9e48-bc4bf74980ab-combined-ca-bundle\") pod \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\" (UID: \"e0e7a440-aafd-4452-9e48-bc4bf74980ab\") " Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.847866 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0e7a440-aafd-4452-9e48-bc4bf74980ab-logs" (OuterVolumeSpecName: "logs") pod "e0e7a440-aafd-4452-9e48-bc4bf74980ab" (UID: "e0e7a440-aafd-4452-9e48-bc4bf74980ab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.848274 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0e7a440-aafd-4452-9e48-bc4bf74980ab-logs\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.856558 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0e7a440-aafd-4452-9e48-bc4bf74980ab-kube-api-access-2hwxt" (OuterVolumeSpecName: "kube-api-access-2hwxt") pod "e0e7a440-aafd-4452-9e48-bc4bf74980ab" (UID: "e0e7a440-aafd-4452-9e48-bc4bf74980ab"). InnerVolumeSpecName "kube-api-access-2hwxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.883055 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0e7a440-aafd-4452-9e48-bc4bf74980ab-config-data" (OuterVolumeSpecName: "config-data") pod "e0e7a440-aafd-4452-9e48-bc4bf74980ab" (UID: "e0e7a440-aafd-4452-9e48-bc4bf74980ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.885234 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0e7a440-aafd-4452-9e48-bc4bf74980ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0e7a440-aafd-4452-9e48-bc4bf74980ab" (UID: "e0e7a440-aafd-4452-9e48-bc4bf74980ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.956874 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e7a440-aafd-4452-9e48-bc4bf74980ab-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.956916 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hwxt\" (UniqueName: \"kubernetes.io/projected/e0e7a440-aafd-4452-9e48-bc4bf74980ab-kube-api-access-2hwxt\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:43 crc kubenswrapper[4744]: I1008 09:30:43.956928 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e7a440-aafd-4452-9e48-bc4bf74980ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.746277 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.822442 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.823039 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.854525 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:44 crc kubenswrapper[4744]: E1008 09:30:44.854974 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0e7a440-aafd-4452-9e48-bc4bf74980ab" containerName="nova-metadata-log" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.854991 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0e7a440-aafd-4452-9e48-bc4bf74980ab" containerName="nova-metadata-log" Oct 08 09:30:44 crc kubenswrapper[4744]: E1008 09:30:44.855017 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0e7a440-aafd-4452-9e48-bc4bf74980ab" containerName="nova-metadata-metadata" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.855024 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0e7a440-aafd-4452-9e48-bc4bf74980ab" containerName="nova-metadata-metadata" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.855212 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0e7a440-aafd-4452-9e48-bc4bf74980ab" containerName="nova-metadata-metadata" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.855228 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0e7a440-aafd-4452-9e48-bc4bf74980ab" containerName="nova-metadata-log" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.860279 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.862790 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.862992 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.872432 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.992180 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.992251 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-config-data\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.992504 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bad7c845-dede-4279-93f2-55da37cfc401-logs\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.992682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj8s9\" (UniqueName: \"kubernetes.io/projected/bad7c845-dede-4279-93f2-55da37cfc401-kube-api-access-zj8s9\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:44 crc kubenswrapper[4744]: I1008 09:30:44.992805 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.094807 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bad7c845-dede-4279-93f2-55da37cfc401-logs\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.094897 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj8s9\" (UniqueName: \"kubernetes.io/projected/bad7c845-dede-4279-93f2-55da37cfc401-kube-api-access-zj8s9\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.094943 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.095060 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.095106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-config-data\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.095276 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bad7c845-dede-4279-93f2-55da37cfc401-logs\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.100817 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.100940 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-config-data\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.101355 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.147214 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj8s9\" (UniqueName: \"kubernetes.io/projected/bad7c845-dede-4279-93f2-55da37cfc401-kube-api-access-zj8s9\") pod \"nova-metadata-0\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " pod="openstack/nova-metadata-0" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.185435 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.467998 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0e7a440-aafd-4452-9e48-bc4bf74980ab" path="/var/lib/kubelet/pods/e0e7a440-aafd-4452-9e48-bc4bf74980ab/volumes" Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.656076 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:45 crc kubenswrapper[4744]: W1008 09:30:45.721234 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbad7c845_dede_4279_93f2_55da37cfc401.slice/crio-e8e700b390502bf4b767dbe458fc25f1c5918d37655472f3f58e45315e00a2d6 WatchSource:0}: Error finding container e8e700b390502bf4b767dbe458fc25f1c5918d37655472f3f58e45315e00a2d6: Status 404 returned error can't find the container with id e8e700b390502bf4b767dbe458fc25f1c5918d37655472f3f58e45315e00a2d6 Oct 08 09:30:45 crc kubenswrapper[4744]: I1008 09:30:45.759704 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bad7c845-dede-4279-93f2-55da37cfc401","Type":"ContainerStarted","Data":"e8e700b390502bf4b767dbe458fc25f1c5918d37655472f3f58e45315e00a2d6"} Oct 08 09:30:46 crc kubenswrapper[4744]: I1008 09:30:46.772265 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bad7c845-dede-4279-93f2-55da37cfc401","Type":"ContainerStarted","Data":"1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280"} Oct 08 09:30:46 crc kubenswrapper[4744]: I1008 09:30:46.774320 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bad7c845-dede-4279-93f2-55da37cfc401","Type":"ContainerStarted","Data":"aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54"} Oct 08 09:30:46 crc kubenswrapper[4744]: I1008 09:30:46.804159 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.804137716 podStartE2EDuration="2.804137716s" podCreationTimestamp="2025-10-08 09:30:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:30:46.802310807 +0000 UTC m=+1142.049956046" watchObservedRunningTime="2025-10-08 09:30:46.804137716 +0000 UTC m=+1142.051782955" Oct 08 09:30:47 crc kubenswrapper[4744]: I1008 09:30:47.029063 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 09:30:47 crc kubenswrapper[4744]: I1008 09:30:47.029118 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 09:30:47 crc kubenswrapper[4744]: I1008 09:30:47.068774 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 09:30:47 crc kubenswrapper[4744]: I1008 09:30:47.087142 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 09:30:47 crc kubenswrapper[4744]: I1008 09:30:47.088109 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 09:30:47 crc kubenswrapper[4744]: I1008 09:30:47.188656 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:30:47 crc kubenswrapper[4744]: I1008 09:30:47.207835 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:30:47 crc kubenswrapper[4744]: I1008 09:30:47.269961 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7"] Oct 08 09:30:47 crc kubenswrapper[4744]: I1008 09:30:47.270238 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" podUID="24751171-e89c-41e2-b441-654ba81d749b" containerName="dnsmasq-dns" containerID="cri-o://9bc3f464205ea70d2f6d5ae1f40e9c86a7bfc698e87ec37cdfb100842802c36f" gracePeriod=10 Oct 08 09:30:47 crc kubenswrapper[4744]: I1008 09:30:47.801991 4744 generic.go:334] "Generic (PLEG): container finished" podID="24751171-e89c-41e2-b441-654ba81d749b" containerID="9bc3f464205ea70d2f6d5ae1f40e9c86a7bfc698e87ec37cdfb100842802c36f" exitCode=0 Oct 08 09:30:47 crc kubenswrapper[4744]: I1008 09:30:47.802422 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" event={"ID":"24751171-e89c-41e2-b441-654ba81d749b","Type":"ContainerDied","Data":"9bc3f464205ea70d2f6d5ae1f40e9c86a7bfc698e87ec37cdfb100842802c36f"} Oct 08 09:30:47 crc kubenswrapper[4744]: I1008 09:30:47.838237 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.023485 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.170587 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.173:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.170596 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.173:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.184832 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-ovsdbserver-nb\") pod \"24751171-e89c-41e2-b441-654ba81d749b\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.184871 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-ovsdbserver-sb\") pod \"24751171-e89c-41e2-b441-654ba81d749b\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.185028 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-dns-svc\") pod \"24751171-e89c-41e2-b441-654ba81d749b\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.185118 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-config\") pod \"24751171-e89c-41e2-b441-654ba81d749b\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.185198 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7ctw\" (UniqueName: \"kubernetes.io/projected/24751171-e89c-41e2-b441-654ba81d749b-kube-api-access-d7ctw\") pod \"24751171-e89c-41e2-b441-654ba81d749b\" (UID: \"24751171-e89c-41e2-b441-654ba81d749b\") " Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.194710 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24751171-e89c-41e2-b441-654ba81d749b-kube-api-access-d7ctw" (OuterVolumeSpecName: "kube-api-access-d7ctw") pod "24751171-e89c-41e2-b441-654ba81d749b" (UID: "24751171-e89c-41e2-b441-654ba81d749b"). InnerVolumeSpecName "kube-api-access-d7ctw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.258486 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-config" (OuterVolumeSpecName: "config") pod "24751171-e89c-41e2-b441-654ba81d749b" (UID: "24751171-e89c-41e2-b441-654ba81d749b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.266875 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "24751171-e89c-41e2-b441-654ba81d749b" (UID: "24751171-e89c-41e2-b441-654ba81d749b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.284057 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "24751171-e89c-41e2-b441-654ba81d749b" (UID: "24751171-e89c-41e2-b441-654ba81d749b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.285410 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "24751171-e89c-41e2-b441-654ba81d749b" (UID: "24751171-e89c-41e2-b441-654ba81d749b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.287131 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.287162 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.287173 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7ctw\" (UniqueName: \"kubernetes.io/projected/24751171-e89c-41e2-b441-654ba81d749b-kube-api-access-d7ctw\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.287186 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.287198 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/24751171-e89c-41e2-b441-654ba81d749b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.817191 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.817183 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7" event={"ID":"24751171-e89c-41e2-b441-654ba81d749b","Type":"ContainerDied","Data":"ca6664161f914fa99ff1e59491979bd7130a74853e8a16f9668cb0b0e7de592b"} Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.817363 4744 scope.go:117] "RemoveContainer" containerID="9bc3f464205ea70d2f6d5ae1f40e9c86a7bfc698e87ec37cdfb100842802c36f" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.823110 4744 generic.go:334] "Generic (PLEG): container finished" podID="f8711d76-061a-4b5a-a2f6-05996808b79d" containerID="2f8b94fa09ec47e8aeb6aab803b6abe46eea297c395a4af7d144259c22f62236" exitCode=0 Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.823675 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x4khv" event={"ID":"f8711d76-061a-4b5a-a2f6-05996808b79d","Type":"ContainerDied","Data":"2f8b94fa09ec47e8aeb6aab803b6abe46eea297c395a4af7d144259c22f62236"} Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.841659 4744 scope.go:117] "RemoveContainer" containerID="322089b1ad280f38bee2713c7ce87b5a3d854fd301692db2a2f307edfc788b52" Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.880455 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7"] Oct 08 09:30:48 crc kubenswrapper[4744]: I1008 09:30:48.886187 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bdc9d6cdc-8rhc7"] Oct 08 09:30:49 crc kubenswrapper[4744]: I1008 09:30:49.475135 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24751171-e89c-41e2-b441-654ba81d749b" path="/var/lib/kubelet/pods/24751171-e89c-41e2-b441-654ba81d749b/volumes" Oct 08 09:30:49 crc kubenswrapper[4744]: I1008 09:30:49.690150 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:30:49 crc kubenswrapper[4744]: I1008 09:30:49.690234 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:30:49 crc kubenswrapper[4744]: I1008 09:30:49.690295 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:30:49 crc kubenswrapper[4744]: I1008 09:30:49.691313 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc794dea8745685aa845397cb02ef8c3c9ae63ea691d5f60cd28e1c653036510"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 09:30:49 crc kubenswrapper[4744]: I1008 09:30:49.691408 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://fc794dea8745685aa845397cb02ef8c3c9ae63ea691d5f60cd28e1c653036510" gracePeriod=600 Oct 08 09:30:49 crc kubenswrapper[4744]: I1008 09:30:49.844859 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="fc794dea8745685aa845397cb02ef8c3c9ae63ea691d5f60cd28e1c653036510" exitCode=0 Oct 08 09:30:49 crc kubenswrapper[4744]: I1008 09:30:49.844930 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"fc794dea8745685aa845397cb02ef8c3c9ae63ea691d5f60cd28e1c653036510"} Oct 08 09:30:49 crc kubenswrapper[4744]: I1008 09:30:49.844972 4744 scope.go:117] "RemoveContainer" containerID="06edcaf88cdb63c2778de4ebf46b1539980361ea9f30176a88a8565819ce43a4" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.185874 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.186298 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.260644 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.325610 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-config-data\") pod \"f8711d76-061a-4b5a-a2f6-05996808b79d\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.325824 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-combined-ca-bundle\") pod \"f8711d76-061a-4b5a-a2f6-05996808b79d\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.325889 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jlm9\" (UniqueName: \"kubernetes.io/projected/f8711d76-061a-4b5a-a2f6-05996808b79d-kube-api-access-2jlm9\") pod \"f8711d76-061a-4b5a-a2f6-05996808b79d\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.325922 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-scripts\") pod \"f8711d76-061a-4b5a-a2f6-05996808b79d\" (UID: \"f8711d76-061a-4b5a-a2f6-05996808b79d\") " Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.352533 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8711d76-061a-4b5a-a2f6-05996808b79d-kube-api-access-2jlm9" (OuterVolumeSpecName: "kube-api-access-2jlm9") pod "f8711d76-061a-4b5a-a2f6-05996808b79d" (UID: "f8711d76-061a-4b5a-a2f6-05996808b79d"). InnerVolumeSpecName "kube-api-access-2jlm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.357337 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8711d76-061a-4b5a-a2f6-05996808b79d" (UID: "f8711d76-061a-4b5a-a2f6-05996808b79d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.364131 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-config-data" (OuterVolumeSpecName: "config-data") pod "f8711d76-061a-4b5a-a2f6-05996808b79d" (UID: "f8711d76-061a-4b5a-a2f6-05996808b79d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.367066 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-scripts" (OuterVolumeSpecName: "scripts") pod "f8711d76-061a-4b5a-a2f6-05996808b79d" (UID: "f8711d76-061a-4b5a-a2f6-05996808b79d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.429126 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.429161 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.429174 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jlm9\" (UniqueName: \"kubernetes.io/projected/f8711d76-061a-4b5a-a2f6-05996808b79d-kube-api-access-2jlm9\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.429322 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8711d76-061a-4b5a-a2f6-05996808b79d-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.885623 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"936d206815e6d287935cf3c047d48aada7c4fce45edc4c83541f33b9de7f8702"} Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.889733 4744 generic.go:334] "Generic (PLEG): container finished" podID="6bc195e6-d464-4837-b6a0-f11b629e3446" containerID="398904171cafba1a279cf46661a2150313ab50fe0b08ce7ff92fbab6592beb76" exitCode=0 Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.889801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vgxfh" event={"ID":"6bc195e6-d464-4837-b6a0-f11b629e3446","Type":"ContainerDied","Data":"398904171cafba1a279cf46661a2150313ab50fe0b08ce7ff92fbab6592beb76"} Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.892156 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x4khv" event={"ID":"f8711d76-061a-4b5a-a2f6-05996808b79d","Type":"ContainerDied","Data":"57cf58d19c2ced2c9ee8e42546e3a7c6a87928d26ad4aeaf0087958768d5e796"} Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.892181 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57cf58d19c2ced2c9ee8e42546e3a7c6a87928d26ad4aeaf0087958768d5e796" Oct 08 09:30:50 crc kubenswrapper[4744]: I1008 09:30:50.892158 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x4khv" Oct 08 09:30:51 crc kubenswrapper[4744]: I1008 09:30:51.024060 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:30:51 crc kubenswrapper[4744]: I1008 09:30:51.024505 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" containerName="nova-api-api" containerID="cri-o://7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc" gracePeriod=30 Oct 08 09:30:51 crc kubenswrapper[4744]: I1008 09:30:51.024350 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" containerName="nova-api-log" containerID="cri-o://d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85" gracePeriod=30 Oct 08 09:30:51 crc kubenswrapper[4744]: I1008 09:30:51.050812 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:30:51 crc kubenswrapper[4744]: I1008 09:30:51.051035 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0a4bbecd-cd23-4615-86fc-dfbffd82ec5a" containerName="nova-scheduler-scheduler" containerID="cri-o://13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594" gracePeriod=30 Oct 08 09:30:51 crc kubenswrapper[4744]: I1008 09:30:51.066334 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:51 crc kubenswrapper[4744]: I1008 09:30:51.903525 4744 generic.go:334] "Generic (PLEG): container finished" podID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" containerID="d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85" exitCode=143 Oct 08 09:30:51 crc kubenswrapper[4744]: I1008 09:30:51.903623 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ab3407e4-f332-4de8-953a-e92ed39c4e7b","Type":"ContainerDied","Data":"d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85"} Oct 08 09:30:51 crc kubenswrapper[4744]: I1008 09:30:51.904288 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bad7c845-dede-4279-93f2-55da37cfc401" containerName="nova-metadata-log" containerID="cri-o://aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54" gracePeriod=30 Oct 08 09:30:51 crc kubenswrapper[4744]: I1008 09:30:51.904290 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bad7c845-dede-4279-93f2-55da37cfc401" containerName="nova-metadata-metadata" containerID="cri-o://1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280" gracePeriod=30 Oct 08 09:30:51 crc kubenswrapper[4744]: I1008 09:30:51.932803 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 09:30:52 crc kubenswrapper[4744]: E1008 09:30:52.039844 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 09:30:52 crc kubenswrapper[4744]: E1008 09:30:52.046448 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 09:30:52 crc kubenswrapper[4744]: E1008 09:30:52.051563 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 09:30:52 crc kubenswrapper[4744]: E1008 09:30:52.051669 4744 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="0a4bbecd-cd23-4615-86fc-dfbffd82ec5a" containerName="nova-scheduler-scheduler" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.284005 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.370118 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-combined-ca-bundle\") pod \"6bc195e6-d464-4837-b6a0-f11b629e3446\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.370437 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-config-data\") pod \"6bc195e6-d464-4837-b6a0-f11b629e3446\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.370550 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm89g\" (UniqueName: \"kubernetes.io/projected/6bc195e6-d464-4837-b6a0-f11b629e3446-kube-api-access-nm89g\") pod \"6bc195e6-d464-4837-b6a0-f11b629e3446\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.370718 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-scripts\") pod \"6bc195e6-d464-4837-b6a0-f11b629e3446\" (UID: \"6bc195e6-d464-4837-b6a0-f11b629e3446\") " Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.377659 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-scripts" (OuterVolumeSpecName: "scripts") pod "6bc195e6-d464-4837-b6a0-f11b629e3446" (UID: "6bc195e6-d464-4837-b6a0-f11b629e3446"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.391759 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bc195e6-d464-4837-b6a0-f11b629e3446-kube-api-access-nm89g" (OuterVolumeSpecName: "kube-api-access-nm89g") pod "6bc195e6-d464-4837-b6a0-f11b629e3446" (UID: "6bc195e6-d464-4837-b6a0-f11b629e3446"). InnerVolumeSpecName "kube-api-access-nm89g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.445866 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bc195e6-d464-4837-b6a0-f11b629e3446" (UID: "6bc195e6-d464-4837-b6a0-f11b629e3446"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.451889 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-config-data" (OuterVolumeSpecName: "config-data") pod "6bc195e6-d464-4837-b6a0-f11b629e3446" (UID: "6bc195e6-d464-4837-b6a0-f11b629e3446"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.474716 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm89g\" (UniqueName: \"kubernetes.io/projected/6bc195e6-d464-4837-b6a0-f11b629e3446-kube-api-access-nm89g\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.474767 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.474781 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.474791 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bc195e6-d464-4837-b6a0-f11b629e3446-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.531755 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.678799 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-nova-metadata-tls-certs\") pod \"bad7c845-dede-4279-93f2-55da37cfc401\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.679323 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj8s9\" (UniqueName: \"kubernetes.io/projected/bad7c845-dede-4279-93f2-55da37cfc401-kube-api-access-zj8s9\") pod \"bad7c845-dede-4279-93f2-55da37cfc401\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.679362 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-config-data\") pod \"bad7c845-dede-4279-93f2-55da37cfc401\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.679433 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bad7c845-dede-4279-93f2-55da37cfc401-logs\") pod \"bad7c845-dede-4279-93f2-55da37cfc401\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.679543 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-combined-ca-bundle\") pod \"bad7c845-dede-4279-93f2-55da37cfc401\" (UID: \"bad7c845-dede-4279-93f2-55da37cfc401\") " Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.679879 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bad7c845-dede-4279-93f2-55da37cfc401-logs" (OuterVolumeSpecName: "logs") pod "bad7c845-dede-4279-93f2-55da37cfc401" (UID: "bad7c845-dede-4279-93f2-55da37cfc401"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.680359 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bad7c845-dede-4279-93f2-55da37cfc401-logs\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.685067 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bad7c845-dede-4279-93f2-55da37cfc401-kube-api-access-zj8s9" (OuterVolumeSpecName: "kube-api-access-zj8s9") pod "bad7c845-dede-4279-93f2-55da37cfc401" (UID: "bad7c845-dede-4279-93f2-55da37cfc401"). InnerVolumeSpecName "kube-api-access-zj8s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.715714 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-config-data" (OuterVolumeSpecName: "config-data") pod "bad7c845-dede-4279-93f2-55da37cfc401" (UID: "bad7c845-dede-4279-93f2-55da37cfc401"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.717475 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bad7c845-dede-4279-93f2-55da37cfc401" (UID: "bad7c845-dede-4279-93f2-55da37cfc401"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.741476 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "bad7c845-dede-4279-93f2-55da37cfc401" (UID: "bad7c845-dede-4279-93f2-55da37cfc401"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.783073 4744 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.783113 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj8s9\" (UniqueName: \"kubernetes.io/projected/bad7c845-dede-4279-93f2-55da37cfc401-kube-api-access-zj8s9\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.783126 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.783138 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad7c845-dede-4279-93f2-55da37cfc401-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.944918 4744 generic.go:334] "Generic (PLEG): container finished" podID="bad7c845-dede-4279-93f2-55da37cfc401" containerID="1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280" exitCode=0 Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.944955 4744 generic.go:334] "Generic (PLEG): container finished" podID="bad7c845-dede-4279-93f2-55da37cfc401" containerID="aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54" exitCode=143 Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.945011 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bad7c845-dede-4279-93f2-55da37cfc401","Type":"ContainerDied","Data":"1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280"} Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.945044 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bad7c845-dede-4279-93f2-55da37cfc401","Type":"ContainerDied","Data":"aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54"} Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.945054 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bad7c845-dede-4279-93f2-55da37cfc401","Type":"ContainerDied","Data":"e8e700b390502bf4b767dbe458fc25f1c5918d37655472f3f58e45315e00a2d6"} Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.945074 4744 scope.go:117] "RemoveContainer" containerID="1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.945226 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.960776 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vgxfh" event={"ID":"6bc195e6-d464-4837-b6a0-f11b629e3446","Type":"ContainerDied","Data":"e8da1283e79e847b1355c6e6ae5f83ddb3e34204776699b105b323166fabea8f"} Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.960838 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8da1283e79e847b1355c6e6ae5f83ddb3e34204776699b105b323166fabea8f" Oct 08 09:30:52 crc kubenswrapper[4744]: I1008 09:30:52.960942 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vgxfh" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.008327 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.028487 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.032556 4744 scope.go:117] "RemoveContainer" containerID="aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.039935 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 09:30:53 crc kubenswrapper[4744]: E1008 09:30:53.040436 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad7c845-dede-4279-93f2-55da37cfc401" containerName="nova-metadata-metadata" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.040449 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad7c845-dede-4279-93f2-55da37cfc401" containerName="nova-metadata-metadata" Oct 08 09:30:53 crc kubenswrapper[4744]: E1008 09:30:53.040465 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24751171-e89c-41e2-b441-654ba81d749b" containerName="dnsmasq-dns" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.040471 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="24751171-e89c-41e2-b441-654ba81d749b" containerName="dnsmasq-dns" Oct 08 09:30:53 crc kubenswrapper[4744]: E1008 09:30:53.040490 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bc195e6-d464-4837-b6a0-f11b629e3446" containerName="nova-cell1-conductor-db-sync" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.040498 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bc195e6-d464-4837-b6a0-f11b629e3446" containerName="nova-cell1-conductor-db-sync" Oct 08 09:30:53 crc kubenswrapper[4744]: E1008 09:30:53.040510 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad7c845-dede-4279-93f2-55da37cfc401" containerName="nova-metadata-log" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.040516 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad7c845-dede-4279-93f2-55da37cfc401" containerName="nova-metadata-log" Oct 08 09:30:53 crc kubenswrapper[4744]: E1008 09:30:53.040530 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8711d76-061a-4b5a-a2f6-05996808b79d" containerName="nova-manage" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.040535 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8711d76-061a-4b5a-a2f6-05996808b79d" containerName="nova-manage" Oct 08 09:30:53 crc kubenswrapper[4744]: E1008 09:30:53.040554 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24751171-e89c-41e2-b441-654ba81d749b" containerName="init" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.040560 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="24751171-e89c-41e2-b441-654ba81d749b" containerName="init" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.040788 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad7c845-dede-4279-93f2-55da37cfc401" containerName="nova-metadata-log" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.040807 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="24751171-e89c-41e2-b441-654ba81d749b" containerName="dnsmasq-dns" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.040814 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8711d76-061a-4b5a-a2f6-05996808b79d" containerName="nova-manage" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.040824 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad7c845-dede-4279-93f2-55da37cfc401" containerName="nova-metadata-metadata" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.040835 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bc195e6-d464-4837-b6a0-f11b629e3446" containerName="nova-cell1-conductor-db-sync" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.041563 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.055290 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.057031 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.057125 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.058118 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.066732 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.067012 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.071692 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.128086 4744 scope.go:117] "RemoveContainer" containerID="1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280" Oct 08 09:30:53 crc kubenswrapper[4744]: E1008 09:30:53.131434 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280\": container with ID starting with 1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280 not found: ID does not exist" containerID="1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.131481 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280"} err="failed to get container status \"1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280\": rpc error: code = NotFound desc = could not find container \"1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280\": container with ID starting with 1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280 not found: ID does not exist" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.131510 4744 scope.go:117] "RemoveContainer" containerID="aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54" Oct 08 09:30:53 crc kubenswrapper[4744]: E1008 09:30:53.131827 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54\": container with ID starting with aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54 not found: ID does not exist" containerID="aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.131844 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54"} err="failed to get container status \"aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54\": rpc error: code = NotFound desc = could not find container \"aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54\": container with ID starting with aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54 not found: ID does not exist" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.131859 4744 scope.go:117] "RemoveContainer" containerID="1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.132505 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280"} err="failed to get container status \"1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280\": rpc error: code = NotFound desc = could not find container \"1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280\": container with ID starting with 1e8c9c88acde5d625333f2ec09d16b5aceb7667e13d656bba516fe4067bfc280 not found: ID does not exist" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.132529 4744 scope.go:117] "RemoveContainer" containerID="aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.132954 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54"} err="failed to get container status \"aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54\": rpc error: code = NotFound desc = could not find container \"aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54\": container with ID starting with aec1b71747fd9cebd0bbcb0758a18002e291449ee2ca5cba16b3bf5530e1dd54 not found: ID does not exist" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.209386 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b03c126-e043-4a2e-ae42-e9a359622fcd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0b03c126-e043-4a2e-ae42-e9a359622fcd\") " pod="openstack/nova-cell1-conductor-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.209562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7f5d\" (UniqueName: \"kubernetes.io/projected/237f9dff-1f92-46fa-a8d3-23f447f334be-kube-api-access-g7f5d\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.209602 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b03c126-e043-4a2e-ae42-e9a359622fcd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0b03c126-e043-4a2e-ae42-e9a359622fcd\") " pod="openstack/nova-cell1-conductor-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.209630 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjnpm\" (UniqueName: \"kubernetes.io/projected/0b03c126-e043-4a2e-ae42-e9a359622fcd-kube-api-access-bjnpm\") pod \"nova-cell1-conductor-0\" (UID: \"0b03c126-e043-4a2e-ae42-e9a359622fcd\") " pod="openstack/nova-cell1-conductor-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.209836 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-config-data\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.209955 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.210096 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/237f9dff-1f92-46fa-a8d3-23f447f334be-logs\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.210220 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.313267 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7f5d\" (UniqueName: \"kubernetes.io/projected/237f9dff-1f92-46fa-a8d3-23f447f334be-kube-api-access-g7f5d\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.313846 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b03c126-e043-4a2e-ae42-e9a359622fcd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0b03c126-e043-4a2e-ae42-e9a359622fcd\") " pod="openstack/nova-cell1-conductor-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.313885 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjnpm\" (UniqueName: \"kubernetes.io/projected/0b03c126-e043-4a2e-ae42-e9a359622fcd-kube-api-access-bjnpm\") pod \"nova-cell1-conductor-0\" (UID: \"0b03c126-e043-4a2e-ae42-e9a359622fcd\") " pod="openstack/nova-cell1-conductor-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.313974 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-config-data\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.314056 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.314141 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/237f9dff-1f92-46fa-a8d3-23f447f334be-logs\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.314231 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.314417 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b03c126-e043-4a2e-ae42-e9a359622fcd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0b03c126-e043-4a2e-ae42-e9a359622fcd\") " pod="openstack/nova-cell1-conductor-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.315603 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/237f9dff-1f92-46fa-a8d3-23f447f334be-logs\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.319479 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.319927 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.321111 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b03c126-e043-4a2e-ae42-e9a359622fcd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0b03c126-e043-4a2e-ae42-e9a359622fcd\") " pod="openstack/nova-cell1-conductor-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.329025 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-config-data\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.332270 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b03c126-e043-4a2e-ae42-e9a359622fcd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0b03c126-e043-4a2e-ae42-e9a359622fcd\") " pod="openstack/nova-cell1-conductor-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.335187 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7f5d\" (UniqueName: \"kubernetes.io/projected/237f9dff-1f92-46fa-a8d3-23f447f334be-kube-api-access-g7f5d\") pod \"nova-metadata-0\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.337913 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjnpm\" (UniqueName: \"kubernetes.io/projected/0b03c126-e043-4a2e-ae42-e9a359622fcd-kube-api-access-bjnpm\") pod \"nova-cell1-conductor-0\" (UID: \"0b03c126-e043-4a2e-ae42-e9a359622fcd\") " pod="openstack/nova-cell1-conductor-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.409959 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.436261 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.466598 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bad7c845-dede-4279-93f2-55da37cfc401" path="/var/lib/kubelet/pods/bad7c845-dede-4279-93f2-55da37cfc401/volumes" Oct 08 09:30:53 crc kubenswrapper[4744]: I1008 09:30:53.964238 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 09:30:53 crc kubenswrapper[4744]: W1008 09:30:53.965586 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b03c126_e043_4a2e_ae42_e9a359622fcd.slice/crio-f522ca78cab28a9c48caa9e1663e2ba444fa73520f9d20c1e2d700669dc07989 WatchSource:0}: Error finding container f522ca78cab28a9c48caa9e1663e2ba444fa73520f9d20c1e2d700669dc07989: Status 404 returned error can't find the container with id f522ca78cab28a9c48caa9e1663e2ba444fa73520f9d20c1e2d700669dc07989 Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.115175 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:30:54 crc kubenswrapper[4744]: W1008 09:30:54.127033 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod237f9dff_1f92_46fa_a8d3_23f447f334be.slice/crio-2cebdb2c9a7405308921569794db4071a1ed0e42d8befef5863153233504dab5 WatchSource:0}: Error finding container 2cebdb2c9a7405308921569794db4071a1ed0e42d8befef5863153233504dab5: Status 404 returned error can't find the container with id 2cebdb2c9a7405308921569794db4071a1ed0e42d8befef5863153233504dab5 Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.629257 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.747585 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab3407e4-f332-4de8-953a-e92ed39c4e7b-logs\") pod \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.747748 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfkds\" (UniqueName: \"kubernetes.io/projected/ab3407e4-f332-4de8-953a-e92ed39c4e7b-kube-api-access-bfkds\") pod \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.747798 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3407e4-f332-4de8-953a-e92ed39c4e7b-combined-ca-bundle\") pod \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.747831 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3407e4-f332-4de8-953a-e92ed39c4e7b-config-data\") pod \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\" (UID: \"ab3407e4-f332-4de8-953a-e92ed39c4e7b\") " Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.748240 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab3407e4-f332-4de8-953a-e92ed39c4e7b-logs" (OuterVolumeSpecName: "logs") pod "ab3407e4-f332-4de8-953a-e92ed39c4e7b" (UID: "ab3407e4-f332-4de8-953a-e92ed39c4e7b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.749091 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab3407e4-f332-4de8-953a-e92ed39c4e7b-logs\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.760203 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab3407e4-f332-4de8-953a-e92ed39c4e7b-kube-api-access-bfkds" (OuterVolumeSpecName: "kube-api-access-bfkds") pod "ab3407e4-f332-4de8-953a-e92ed39c4e7b" (UID: "ab3407e4-f332-4de8-953a-e92ed39c4e7b"). InnerVolumeSpecName "kube-api-access-bfkds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.798491 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3407e4-f332-4de8-953a-e92ed39c4e7b-config-data" (OuterVolumeSpecName: "config-data") pod "ab3407e4-f332-4de8-953a-e92ed39c4e7b" (UID: "ab3407e4-f332-4de8-953a-e92ed39c4e7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.841695 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3407e4-f332-4de8-953a-e92ed39c4e7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab3407e4-f332-4de8-953a-e92ed39c4e7b" (UID: "ab3407e4-f332-4de8-953a-e92ed39c4e7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.867151 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfkds\" (UniqueName: \"kubernetes.io/projected/ab3407e4-f332-4de8-953a-e92ed39c4e7b-kube-api-access-bfkds\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.867183 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3407e4-f332-4de8-953a-e92ed39c4e7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.867197 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab3407e4-f332-4de8-953a-e92ed39c4e7b-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.985578 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.993713 4744 generic.go:334] "Generic (PLEG): container finished" podID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" containerID="7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc" exitCode=0 Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.993779 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ab3407e4-f332-4de8-953a-e92ed39c4e7b","Type":"ContainerDied","Data":"7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc"} Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.993809 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ab3407e4-f332-4de8-953a-e92ed39c4e7b","Type":"ContainerDied","Data":"c14d146e6f86f7e8c76477c5941c5a6390f1ceb9010acc9239adbfe508ecbb54"} Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.993830 4744 scope.go:117] "RemoveContainer" containerID="7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc" Oct 08 09:30:54 crc kubenswrapper[4744]: I1008 09:30:54.993970 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.032724 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"237f9dff-1f92-46fa-a8d3-23f447f334be","Type":"ContainerStarted","Data":"7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8"} Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.033114 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"237f9dff-1f92-46fa-a8d3-23f447f334be","Type":"ContainerStarted","Data":"0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21"} Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.033574 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"237f9dff-1f92-46fa-a8d3-23f447f334be","Type":"ContainerStarted","Data":"2cebdb2c9a7405308921569794db4071a1ed0e42d8befef5863153233504dab5"} Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.059096 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0b03c126-e043-4a2e-ae42-e9a359622fcd","Type":"ContainerStarted","Data":"b37ff412ac1b7cdcd61ac215279d84bf0447db0a9e8d1f967c01b416e4a51f47"} Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.059148 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0b03c126-e043-4a2e-ae42-e9a359622fcd","Type":"ContainerStarted","Data":"f522ca78cab28a9c48caa9e1663e2ba444fa73520f9d20c1e2d700669dc07989"} Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.059564 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.102843 4744 generic.go:334] "Generic (PLEG): container finished" podID="0a4bbecd-cd23-4615-86fc-dfbffd82ec5a" containerID="13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594" exitCode=0 Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.102889 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a","Type":"ContainerDied","Data":"13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594"} Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.102920 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a","Type":"ContainerDied","Data":"82154c8240c2c8028252194c92e3ca26098a76d05736d01cf9e67016932e4f20"} Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.102976 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.112133 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.112088058 podStartE2EDuration="2.112088058s" podCreationTimestamp="2025-10-08 09:30:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:30:55.074928085 +0000 UTC m=+1150.322573324" watchObservedRunningTime="2025-10-08 09:30:55.112088058 +0000 UTC m=+1150.359733297" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.136588 4744 scope.go:117] "RemoveContainer" containerID="d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.137551 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.137517648 podStartE2EDuration="3.137517648s" podCreationTimestamp="2025-10-08 09:30:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:30:55.111625986 +0000 UTC m=+1150.359271225" watchObservedRunningTime="2025-10-08 09:30:55.137517648 +0000 UTC m=+1150.385162887" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.175704 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-combined-ca-bundle\") pod \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\" (UID: \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\") " Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.177740 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-config-data\") pod \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\" (UID: \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\") " Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.177771 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqkvs\" (UniqueName: \"kubernetes.io/projected/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-kube-api-access-nqkvs\") pod \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\" (UID: \"0a4bbecd-cd23-4615-86fc-dfbffd82ec5a\") " Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.191673 4744 scope.go:117] "RemoveContainer" containerID="7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc" Oct 08 09:30:55 crc kubenswrapper[4744]: E1008 09:30:55.200567 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc\": container with ID starting with 7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc not found: ID does not exist" containerID="7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.200625 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc"} err="failed to get container status \"7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc\": rpc error: code = NotFound desc = could not find container \"7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc\": container with ID starting with 7b67ba2d3c2393d1311a286586e542241df57b7c2678671a1bbf782bbf371bcc not found: ID does not exist" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.200655 4744 scope.go:117] "RemoveContainer" containerID="d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.200922 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-kube-api-access-nqkvs" (OuterVolumeSpecName: "kube-api-access-nqkvs") pod "0a4bbecd-cd23-4615-86fc-dfbffd82ec5a" (UID: "0a4bbecd-cd23-4615-86fc-dfbffd82ec5a"). InnerVolumeSpecName "kube-api-access-nqkvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:55 crc kubenswrapper[4744]: E1008 09:30:55.208680 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85\": container with ID starting with d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85 not found: ID does not exist" containerID="d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.208752 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85"} err="failed to get container status \"d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85\": rpc error: code = NotFound desc = could not find container \"d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85\": container with ID starting with d008fbbaf49dcd89b4082dd502495214635c1cf81d133aaee538dda42bbc3a85 not found: ID does not exist" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.208809 4744 scope.go:117] "RemoveContainer" containerID="13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.229227 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.277631 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.294767 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqkvs\" (UniqueName: \"kubernetes.io/projected/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-kube-api-access-nqkvs\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.305848 4744 scope.go:117] "RemoveContainer" containerID="13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594" Oct 08 09:30:55 crc kubenswrapper[4744]: E1008 09:30:55.332766 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594\": container with ID starting with 13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594 not found: ID does not exist" containerID="13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.332817 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594"} err="failed to get container status \"13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594\": rpc error: code = NotFound desc = could not find container \"13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594\": container with ID starting with 13bae4c1f79f8bc649289c43f284c180b92d65e91e4047455529c01cf490a594 not found: ID does not exist" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.338748 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-config-data" (OuterVolumeSpecName: "config-data") pod "0a4bbecd-cd23-4615-86fc-dfbffd82ec5a" (UID: "0a4bbecd-cd23-4615-86fc-dfbffd82ec5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.339031 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a4bbecd-cd23-4615-86fc-dfbffd82ec5a" (UID: "0a4bbecd-cd23-4615-86fc-dfbffd82ec5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.382499 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 09:30:55 crc kubenswrapper[4744]: E1008 09:30:55.383382 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" containerName="nova-api-log" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.383407 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" containerName="nova-api-log" Oct 08 09:30:55 crc kubenswrapper[4744]: E1008 09:30:55.383449 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4bbecd-cd23-4615-86fc-dfbffd82ec5a" containerName="nova-scheduler-scheduler" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.383456 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4bbecd-cd23-4615-86fc-dfbffd82ec5a" containerName="nova-scheduler-scheduler" Oct 08 09:30:55 crc kubenswrapper[4744]: E1008 09:30:55.383899 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" containerName="nova-api-api" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.383914 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" containerName="nova-api-api" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.384370 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" containerName="nova-api-api" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.384445 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a4bbecd-cd23-4615-86fc-dfbffd82ec5a" containerName="nova-scheduler-scheduler" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.384459 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" containerName="nova-api-log" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.386148 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.390116 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.397776 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.397795 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.445686 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.494303 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab3407e4-f332-4de8-953a-e92ed39c4e7b" path="/var/lib/kubelet/pods/ab3407e4-f332-4de8-953a-e92ed39c4e7b/volumes" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.495368 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.495412 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.501550 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729a216f-b9ac-400e-bae4-c3cfe7e4023d-config-data\") pod \"nova-api-0\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.501609 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxxms\" (UniqueName: \"kubernetes.io/projected/729a216f-b9ac-400e-bae4-c3cfe7e4023d-kube-api-access-nxxms\") pod \"nova-api-0\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.501726 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/729a216f-b9ac-400e-bae4-c3cfe7e4023d-logs\") pod \"nova-api-0\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.501798 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729a216f-b9ac-400e-bae4-c3cfe7e4023d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.540603 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.552744 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.567342 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.574132 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.603312 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/729a216f-b9ac-400e-bae4-c3cfe7e4023d-logs\") pod \"nova-api-0\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.603410 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729a216f-b9ac-400e-bae4-c3cfe7e4023d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.603455 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729a216f-b9ac-400e-bae4-c3cfe7e4023d-config-data\") pod \"nova-api-0\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.603488 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxxms\" (UniqueName: \"kubernetes.io/projected/729a216f-b9ac-400e-bae4-c3cfe7e4023d-kube-api-access-nxxms\") pod \"nova-api-0\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.605046 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/729a216f-b9ac-400e-bae4-c3cfe7e4023d-logs\") pod \"nova-api-0\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.610812 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729a216f-b9ac-400e-bae4-c3cfe7e4023d-config-data\") pod \"nova-api-0\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.616170 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729a216f-b9ac-400e-bae4-c3cfe7e4023d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.629747 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxxms\" (UniqueName: \"kubernetes.io/projected/729a216f-b9ac-400e-bae4-c3cfe7e4023d-kube-api-access-nxxms\") pod \"nova-api-0\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.702029 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.705073 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8652l\" (UniqueName: \"kubernetes.io/projected/4e3d9501-0500-4fea-a909-0d3845a1845d-kube-api-access-8652l\") pod \"nova-scheduler-0\" (UID: \"4e3d9501-0500-4fea-a909-0d3845a1845d\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.705291 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e3d9501-0500-4fea-a909-0d3845a1845d-config-data\") pod \"nova-scheduler-0\" (UID: \"4e3d9501-0500-4fea-a909-0d3845a1845d\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.705340 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e3d9501-0500-4fea-a909-0d3845a1845d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4e3d9501-0500-4fea-a909-0d3845a1845d\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.749290 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.749587 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="7eb871ad-c10b-48ea-b27e-f9dfca634845" containerName="kube-state-metrics" containerID="cri-o://54d39aa20d340d7dce09a9f85db69d33883b237537de14c0bde01b8fc28ad6f7" gracePeriod=30 Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.806542 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8652l\" (UniqueName: \"kubernetes.io/projected/4e3d9501-0500-4fea-a909-0d3845a1845d-kube-api-access-8652l\") pod \"nova-scheduler-0\" (UID: \"4e3d9501-0500-4fea-a909-0d3845a1845d\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.806682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e3d9501-0500-4fea-a909-0d3845a1845d-config-data\") pod \"nova-scheduler-0\" (UID: \"4e3d9501-0500-4fea-a909-0d3845a1845d\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.806712 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e3d9501-0500-4fea-a909-0d3845a1845d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4e3d9501-0500-4fea-a909-0d3845a1845d\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.811173 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e3d9501-0500-4fea-a909-0d3845a1845d-config-data\") pod \"nova-scheduler-0\" (UID: \"4e3d9501-0500-4fea-a909-0d3845a1845d\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.811363 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e3d9501-0500-4fea-a909-0d3845a1845d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4e3d9501-0500-4fea-a909-0d3845a1845d\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.827972 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8652l\" (UniqueName: \"kubernetes.io/projected/4e3d9501-0500-4fea-a909-0d3845a1845d-kube-api-access-8652l\") pod \"nova-scheduler-0\" (UID: \"4e3d9501-0500-4fea-a909-0d3845a1845d\") " pod="openstack/nova-scheduler-0" Oct 08 09:30:55 crc kubenswrapper[4744]: I1008 09:30:55.903136 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 09:30:56 crc kubenswrapper[4744]: I1008 09:30:56.129451 4744 generic.go:334] "Generic (PLEG): container finished" podID="7eb871ad-c10b-48ea-b27e-f9dfca634845" containerID="54d39aa20d340d7dce09a9f85db69d33883b237537de14c0bde01b8fc28ad6f7" exitCode=2 Oct 08 09:30:56 crc kubenswrapper[4744]: I1008 09:30:56.130631 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7eb871ad-c10b-48ea-b27e-f9dfca634845","Type":"ContainerDied","Data":"54d39aa20d340d7dce09a9f85db69d33883b237537de14c0bde01b8fc28ad6f7"} Oct 08 09:30:56 crc kubenswrapper[4744]: I1008 09:30:56.295050 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:30:56 crc kubenswrapper[4744]: I1008 09:30:56.442301 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 09:30:56 crc kubenswrapper[4744]: I1008 09:30:56.526563 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcm2r\" (UniqueName: \"kubernetes.io/projected/7eb871ad-c10b-48ea-b27e-f9dfca634845-kube-api-access-pcm2r\") pod \"7eb871ad-c10b-48ea-b27e-f9dfca634845\" (UID: \"7eb871ad-c10b-48ea-b27e-f9dfca634845\") " Oct 08 09:30:56 crc kubenswrapper[4744]: I1008 09:30:56.561123 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eb871ad-c10b-48ea-b27e-f9dfca634845-kube-api-access-pcm2r" (OuterVolumeSpecName: "kube-api-access-pcm2r") pod "7eb871ad-c10b-48ea-b27e-f9dfca634845" (UID: "7eb871ad-c10b-48ea-b27e-f9dfca634845"). InnerVolumeSpecName "kube-api-access-pcm2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:30:56 crc kubenswrapper[4744]: I1008 09:30:56.629899 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcm2r\" (UniqueName: \"kubernetes.io/projected/7eb871ad-c10b-48ea-b27e-f9dfca634845-kube-api-access-pcm2r\") on node \"crc\" DevicePath \"\"" Oct 08 09:30:56 crc kubenswrapper[4744]: I1008 09:30:56.719554 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.145451 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7eb871ad-c10b-48ea-b27e-f9dfca634845","Type":"ContainerDied","Data":"a6feff7f60e80b53d447194024690ed34e47c66c7fd7be0df08ede3091ae3754"} Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.145516 4744 scope.go:117] "RemoveContainer" containerID="54d39aa20d340d7dce09a9f85db69d33883b237537de14c0bde01b8fc28ad6f7" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.145646 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.153043 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4e3d9501-0500-4fea-a909-0d3845a1845d","Type":"ContainerStarted","Data":"c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f"} Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.153193 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4e3d9501-0500-4fea-a909-0d3845a1845d","Type":"ContainerStarted","Data":"c34d179eb17554321ae770727d1a04eb5809363bd8d29d43d9ea7f07e84731e8"} Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.157323 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"729a216f-b9ac-400e-bae4-c3cfe7e4023d","Type":"ContainerStarted","Data":"cffa31b9a562fe8c7bfb9599d87bf1723bf56195baef526033907924db285758"} Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.157350 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"729a216f-b9ac-400e-bae4-c3cfe7e4023d","Type":"ContainerStarted","Data":"ea33cb02115338e3fabbb430fe5159429bf7ce934db51851c504a0c1ceca3574"} Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.157360 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"729a216f-b9ac-400e-bae4-c3cfe7e4023d","Type":"ContainerStarted","Data":"cecf06170045e24821547bfb8c4db488244085a09e7dd80427ad76ec2e211965"} Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.179450 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.179425913 podStartE2EDuration="2.179425913s" podCreationTimestamp="2025-10-08 09:30:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:30:57.17108545 +0000 UTC m=+1152.418730689" watchObservedRunningTime="2025-10-08 09:30:57.179425913 +0000 UTC m=+1152.427071162" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.206473 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.229467 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.257791 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 09:30:57 crc kubenswrapper[4744]: E1008 09:30:57.258332 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eb871ad-c10b-48ea-b27e-f9dfca634845" containerName="kube-state-metrics" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.258351 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eb871ad-c10b-48ea-b27e-f9dfca634845" containerName="kube-state-metrics" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.258586 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eb871ad-c10b-48ea-b27e-f9dfca634845" containerName="kube-state-metrics" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.259210 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.263171 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.263597 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.287084 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.287053081 podStartE2EDuration="2.287053081s" podCreationTimestamp="2025-10-08 09:30:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:30:57.228931197 +0000 UTC m=+1152.476576436" watchObservedRunningTime="2025-10-08 09:30:57.287053081 +0000 UTC m=+1152.534698320" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.340594 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.348154 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8b493e3-c187-4150-95db-3a4f110d15ea-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a8b493e3-c187-4150-95db-3a4f110d15ea\") " pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.348290 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b493e3-c187-4150-95db-3a4f110d15ea-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a8b493e3-c187-4150-95db-3a4f110d15ea\") " pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.348439 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a8b493e3-c187-4150-95db-3a4f110d15ea-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a8b493e3-c187-4150-95db-3a4f110d15ea\") " pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.348560 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c2tv\" (UniqueName: \"kubernetes.io/projected/a8b493e3-c187-4150-95db-3a4f110d15ea-kube-api-access-8c2tv\") pod \"kube-state-metrics-0\" (UID: \"a8b493e3-c187-4150-95db-3a4f110d15ea\") " pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.442348 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.443313 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="sg-core" containerID="cri-o://f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33" gracePeriod=30 Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.443504 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="proxy-httpd" containerID="cri-o://bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6" gracePeriod=30 Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.443742 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="ceilometer-notification-agent" containerID="cri-o://217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92" gracePeriod=30 Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.443804 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="ceilometer-central-agent" containerID="cri-o://bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28" gracePeriod=30 Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.450004 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a8b493e3-c187-4150-95db-3a4f110d15ea-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a8b493e3-c187-4150-95db-3a4f110d15ea\") " pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.450606 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c2tv\" (UniqueName: \"kubernetes.io/projected/a8b493e3-c187-4150-95db-3a4f110d15ea-kube-api-access-8c2tv\") pod \"kube-state-metrics-0\" (UID: \"a8b493e3-c187-4150-95db-3a4f110d15ea\") " pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.450664 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8b493e3-c187-4150-95db-3a4f110d15ea-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a8b493e3-c187-4150-95db-3a4f110d15ea\") " pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.450713 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b493e3-c187-4150-95db-3a4f110d15ea-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a8b493e3-c187-4150-95db-3a4f110d15ea\") " pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.482780 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a4bbecd-cd23-4615-86fc-dfbffd82ec5a" path="/var/lib/kubelet/pods/0a4bbecd-cd23-4615-86fc-dfbffd82ec5a/volumes" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.483308 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7eb871ad-c10b-48ea-b27e-f9dfca634845" path="/var/lib/kubelet/pods/7eb871ad-c10b-48ea-b27e-f9dfca634845/volumes" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.486424 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8b493e3-c187-4150-95db-3a4f110d15ea-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a8b493e3-c187-4150-95db-3a4f110d15ea\") " pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.493496 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b493e3-c187-4150-95db-3a4f110d15ea-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a8b493e3-c187-4150-95db-3a4f110d15ea\") " pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.505269 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c2tv\" (UniqueName: \"kubernetes.io/projected/a8b493e3-c187-4150-95db-3a4f110d15ea-kube-api-access-8c2tv\") pod \"kube-state-metrics-0\" (UID: \"a8b493e3-c187-4150-95db-3a4f110d15ea\") " pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.508531 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a8b493e3-c187-4150-95db-3a4f110d15ea-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a8b493e3-c187-4150-95db-3a4f110d15ea\") " pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: I1008 09:30:57.598523 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 09:30:57 crc kubenswrapper[4744]: E1008 09:30:57.636939 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7eb871ad_c10b_48ea_b27e_f9dfca634845.slice/crio-a6feff7f60e80b53d447194024690ed34e47c66c7fd7be0df08ede3091ae3754\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7eb871ad_c10b_48ea_b27e_f9dfca634845.slice\": RecentStats: unable to find data in memory cache]" Oct 08 09:30:58 crc kubenswrapper[4744]: I1008 09:30:58.064833 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 09:30:58 crc kubenswrapper[4744]: I1008 09:30:58.181564 4744 generic.go:334] "Generic (PLEG): container finished" podID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerID="bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6" exitCode=0 Oct 08 09:30:58 crc kubenswrapper[4744]: I1008 09:30:58.181614 4744 generic.go:334] "Generic (PLEG): container finished" podID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerID="f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33" exitCode=2 Oct 08 09:30:58 crc kubenswrapper[4744]: I1008 09:30:58.181667 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b2d9e7-2617-4c74-b4cd-ee619204ee07","Type":"ContainerDied","Data":"bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6"} Oct 08 09:30:58 crc kubenswrapper[4744]: I1008 09:30:58.181695 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b2d9e7-2617-4c74-b4cd-ee619204ee07","Type":"ContainerDied","Data":"f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33"} Oct 08 09:30:58 crc kubenswrapper[4744]: I1008 09:30:58.185986 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a8b493e3-c187-4150-95db-3a4f110d15ea","Type":"ContainerStarted","Data":"183e2cf3bd07f06d0e1eaf0a4df7c57ff5b4c352d198a63978f79db0c5c2f344"} Oct 08 09:30:58 crc kubenswrapper[4744]: I1008 09:30:58.437018 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 09:30:58 crc kubenswrapper[4744]: I1008 09:30:58.437379 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 09:30:59 crc kubenswrapper[4744]: I1008 09:30:59.196461 4744 generic.go:334] "Generic (PLEG): container finished" podID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerID="bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28" exitCode=0 Oct 08 09:30:59 crc kubenswrapper[4744]: I1008 09:30:59.196533 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b2d9e7-2617-4c74-b4cd-ee619204ee07","Type":"ContainerDied","Data":"bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28"} Oct 08 09:30:59 crc kubenswrapper[4744]: I1008 09:30:59.199974 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a8b493e3-c187-4150-95db-3a4f110d15ea","Type":"ContainerStarted","Data":"52ba92b943f5777cad903a3f08a40257058af6331b62f8e7e8c6cbb1aab32501"} Oct 08 09:30:59 crc kubenswrapper[4744]: I1008 09:30:59.200044 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 09:30:59 crc kubenswrapper[4744]: I1008 09:30:59.224763 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.550796363 podStartE2EDuration="2.224740909s" podCreationTimestamp="2025-10-08 09:30:57 +0000 UTC" firstStartedPulling="2025-10-08 09:30:58.094054234 +0000 UTC m=+1153.341699473" lastFinishedPulling="2025-10-08 09:30:58.76799878 +0000 UTC m=+1154.015644019" observedRunningTime="2025-10-08 09:30:59.218296968 +0000 UTC m=+1154.465942227" watchObservedRunningTime="2025-10-08 09:30:59.224740909 +0000 UTC m=+1154.472386148" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.739644 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.838191 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b2d9e7-2617-4c74-b4cd-ee619204ee07-run-httpd\") pod \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.838257 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-sg-core-conf-yaml\") pod \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.838298 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-combined-ca-bundle\") pod \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.838326 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pjqw\" (UniqueName: \"kubernetes.io/projected/68b2d9e7-2617-4c74-b4cd-ee619204ee07-kube-api-access-9pjqw\") pod \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.838387 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-config-data\") pod \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.838432 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-scripts\") pod \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.838476 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b2d9e7-2617-4c74-b4cd-ee619204ee07-log-httpd\") pod \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\" (UID: \"68b2d9e7-2617-4c74-b4cd-ee619204ee07\") " Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.839755 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68b2d9e7-2617-4c74-b4cd-ee619204ee07-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "68b2d9e7-2617-4c74-b4cd-ee619204ee07" (UID: "68b2d9e7-2617-4c74-b4cd-ee619204ee07"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.839928 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68b2d9e7-2617-4c74-b4cd-ee619204ee07-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "68b2d9e7-2617-4c74-b4cd-ee619204ee07" (UID: "68b2d9e7-2617-4c74-b4cd-ee619204ee07"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.850138 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-scripts" (OuterVolumeSpecName: "scripts") pod "68b2d9e7-2617-4c74-b4cd-ee619204ee07" (UID: "68b2d9e7-2617-4c74-b4cd-ee619204ee07"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.864691 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68b2d9e7-2617-4c74-b4cd-ee619204ee07-kube-api-access-9pjqw" (OuterVolumeSpecName: "kube-api-access-9pjqw") pod "68b2d9e7-2617-4c74-b4cd-ee619204ee07" (UID: "68b2d9e7-2617-4c74-b4cd-ee619204ee07"). InnerVolumeSpecName "kube-api-access-9pjqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.903833 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.924462 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "68b2d9e7-2617-4c74-b4cd-ee619204ee07" (UID: "68b2d9e7-2617-4c74-b4cd-ee619204ee07"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.941013 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.941051 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b2d9e7-2617-4c74-b4cd-ee619204ee07-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.941064 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68b2d9e7-2617-4c74-b4cd-ee619204ee07-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.941077 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.941090 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pjqw\" (UniqueName: \"kubernetes.io/projected/68b2d9e7-2617-4c74-b4cd-ee619204ee07-kube-api-access-9pjqw\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.952102 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68b2d9e7-2617-4c74-b4cd-ee619204ee07" (UID: "68b2d9e7-2617-4c74-b4cd-ee619204ee07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:00 crc kubenswrapper[4744]: I1008 09:31:00.982172 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-config-data" (OuterVolumeSpecName: "config-data") pod "68b2d9e7-2617-4c74-b4cd-ee619204ee07" (UID: "68b2d9e7-2617-4c74-b4cd-ee619204ee07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.042294 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.042352 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68b2d9e7-2617-4c74-b4cd-ee619204ee07-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.227479 4744 generic.go:334] "Generic (PLEG): container finished" podID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerID="217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92" exitCode=0 Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.227542 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b2d9e7-2617-4c74-b4cd-ee619204ee07","Type":"ContainerDied","Data":"217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92"} Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.227586 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68b2d9e7-2617-4c74-b4cd-ee619204ee07","Type":"ContainerDied","Data":"7f6979499ef75886cf5ba2776c757ad96ce2a025e88276cc6dd1555beea6f433"} Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.227607 4744 scope.go:117] "RemoveContainer" containerID="bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.227713 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.274662 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.284339 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.286737 4744 scope.go:117] "RemoveContainer" containerID="f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.312861 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:31:01 crc kubenswrapper[4744]: E1008 09:31:01.313423 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="proxy-httpd" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.313441 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="proxy-httpd" Oct 08 09:31:01 crc kubenswrapper[4744]: E1008 09:31:01.313479 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="ceilometer-central-agent" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.313485 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="ceilometer-central-agent" Oct 08 09:31:01 crc kubenswrapper[4744]: E1008 09:31:01.313506 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="sg-core" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.313512 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="sg-core" Oct 08 09:31:01 crc kubenswrapper[4744]: E1008 09:31:01.313528 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="ceilometer-notification-agent" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.313535 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="ceilometer-notification-agent" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.313709 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="ceilometer-central-agent" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.313727 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="sg-core" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.313744 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="ceilometer-notification-agent" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.313760 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" containerName="proxy-httpd" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.315776 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.320782 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.320960 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.321069 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.327852 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.345999 4744 scope.go:117] "RemoveContainer" containerID="217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.353580 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.353682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-scripts\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.353707 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-run-httpd\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.354004 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.355094 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.355208 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rznf8\" (UniqueName: \"kubernetes.io/projected/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-kube-api-access-rznf8\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.355282 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-log-httpd\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.355405 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-config-data\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.377645 4744 scope.go:117] "RemoveContainer" containerID="bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.406520 4744 scope.go:117] "RemoveContainer" containerID="bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6" Oct 08 09:31:01 crc kubenswrapper[4744]: E1008 09:31:01.406986 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6\": container with ID starting with bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6 not found: ID does not exist" containerID="bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.407017 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6"} err="failed to get container status \"bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6\": rpc error: code = NotFound desc = could not find container \"bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6\": container with ID starting with bcf19f448ec337d66f710f4ed40bc6b559f52e53b7e68c145b8960835935e0d6 not found: ID does not exist" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.407045 4744 scope.go:117] "RemoveContainer" containerID="f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33" Oct 08 09:31:01 crc kubenswrapper[4744]: E1008 09:31:01.407981 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33\": container with ID starting with f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33 not found: ID does not exist" containerID="f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.408053 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33"} err="failed to get container status \"f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33\": rpc error: code = NotFound desc = could not find container \"f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33\": container with ID starting with f1a0b4783ee86edd27791364d44a2feba90d59daede5bfe0055e170bb69bda33 not found: ID does not exist" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.408138 4744 scope.go:117] "RemoveContainer" containerID="217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92" Oct 08 09:31:01 crc kubenswrapper[4744]: E1008 09:31:01.408822 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92\": container with ID starting with 217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92 not found: ID does not exist" containerID="217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.408898 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92"} err="failed to get container status \"217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92\": rpc error: code = NotFound desc = could not find container \"217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92\": container with ID starting with 217c6aea66e2aed846960c3a17bd4b0f4d939b2cba7dce863b24f96bffcd6a92 not found: ID does not exist" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.408936 4744 scope.go:117] "RemoveContainer" containerID="bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28" Oct 08 09:31:01 crc kubenswrapper[4744]: E1008 09:31:01.409554 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28\": container with ID starting with bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28 not found: ID does not exist" containerID="bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.409593 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28"} err="failed to get container status \"bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28\": rpc error: code = NotFound desc = could not find container \"bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28\": container with ID starting with bc1506dedb1028a6e6e722366815d02854b92e18b82465ea336594fc52a5ea28 not found: ID does not exist" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.457569 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.457622 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rznf8\" (UniqueName: \"kubernetes.io/projected/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-kube-api-access-rznf8\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.457652 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-log-httpd\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.457677 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-config-data\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.457697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.457754 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-scripts\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.457797 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-run-httpd\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.457824 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.464733 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.464844 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-log-httpd\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.465025 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-run-httpd\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.466214 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.467907 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-scripts\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.468638 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.469247 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-config-data\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.476119 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68b2d9e7-2617-4c74-b4cd-ee619204ee07" path="/var/lib/kubelet/pods/68b2d9e7-2617-4c74-b4cd-ee619204ee07/volumes" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.488805 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rznf8\" (UniqueName: \"kubernetes.io/projected/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-kube-api-access-rznf8\") pod \"ceilometer-0\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " pod="openstack/ceilometer-0" Oct 08 09:31:01 crc kubenswrapper[4744]: I1008 09:31:01.660182 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:31:02 crc kubenswrapper[4744]: I1008 09:31:02.161264 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:31:02 crc kubenswrapper[4744]: W1008 09:31:02.162609 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod188a0d94_bf8c_44cc_bbe2_8cd1d0e75103.slice/crio-f62efb4200d9c821b858b7a296271a5c232a3099e87a6fcc3838d716d5c323d3 WatchSource:0}: Error finding container f62efb4200d9c821b858b7a296271a5c232a3099e87a6fcc3838d716d5c323d3: Status 404 returned error can't find the container with id f62efb4200d9c821b858b7a296271a5c232a3099e87a6fcc3838d716d5c323d3 Oct 08 09:31:02 crc kubenswrapper[4744]: I1008 09:31:02.239987 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103","Type":"ContainerStarted","Data":"f62efb4200d9c821b858b7a296271a5c232a3099e87a6fcc3838d716d5c323d3"} Oct 08 09:31:03 crc kubenswrapper[4744]: I1008 09:31:03.274603 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103","Type":"ContainerStarted","Data":"cf1131054f4311986d504bba63174293e5bcdd172e821fe26b8ccef862962f8a"} Oct 08 09:31:03 crc kubenswrapper[4744]: I1008 09:31:03.437255 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 09:31:03 crc kubenswrapper[4744]: I1008 09:31:03.437321 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 09:31:03 crc kubenswrapper[4744]: I1008 09:31:03.446254 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 08 09:31:04 crc kubenswrapper[4744]: I1008 09:31:04.287981 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103","Type":"ContainerStarted","Data":"15145ab3881f5586c9b41f8450412a1a93c9768f6c966b73cf80d81c5dd34897"} Oct 08 09:31:04 crc kubenswrapper[4744]: I1008 09:31:04.288352 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103","Type":"ContainerStarted","Data":"c1695999861d43b9a189dd373e22e1ce44420c20beb3d3b721db7a2f53faed1a"} Oct 08 09:31:04 crc kubenswrapper[4744]: I1008 09:31:04.452588 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.180:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 09:31:04 crc kubenswrapper[4744]: I1008 09:31:04.453007 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.180:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 09:31:05 crc kubenswrapper[4744]: I1008 09:31:05.704473 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 09:31:05 crc kubenswrapper[4744]: I1008 09:31:05.706295 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 09:31:05 crc kubenswrapper[4744]: I1008 09:31:05.904262 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 09:31:05 crc kubenswrapper[4744]: I1008 09:31:05.946696 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 09:31:06 crc kubenswrapper[4744]: I1008 09:31:06.316001 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103","Type":"ContainerStarted","Data":"df810bf0265cc387e02f67b03de4581a466bbd1ff5cce2b24ecb5ab373fcfdc3"} Oct 08 09:31:06 crc kubenswrapper[4744]: I1008 09:31:06.316151 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 09:31:06 crc kubenswrapper[4744]: I1008 09:31:06.351167 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.392359539 podStartE2EDuration="5.351143126s" podCreationTimestamp="2025-10-08 09:31:01 +0000 UTC" firstStartedPulling="2025-10-08 09:31:02.166171871 +0000 UTC m=+1157.413817110" lastFinishedPulling="2025-10-08 09:31:05.124955458 +0000 UTC m=+1160.372600697" observedRunningTime="2025-10-08 09:31:06.350386706 +0000 UTC m=+1161.598031975" watchObservedRunningTime="2025-10-08 09:31:06.351143126 +0000 UTC m=+1161.598788365" Oct 08 09:31:06 crc kubenswrapper[4744]: I1008 09:31:06.364303 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 09:31:06 crc kubenswrapper[4744]: I1008 09:31:06.786651 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.181:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 09:31:06 crc kubenswrapper[4744]: I1008 09:31:06.786677 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.181:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 09:31:07 crc kubenswrapper[4744]: I1008 09:31:07.617676 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.162116 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.361508 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6156b6be-400b-45e6-afee-7fc443b2b23d-combined-ca-bundle\") pod \"6156b6be-400b-45e6-afee-7fc443b2b23d\" (UID: \"6156b6be-400b-45e6-afee-7fc443b2b23d\") " Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.361743 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6156b6be-400b-45e6-afee-7fc443b2b23d-config-data\") pod \"6156b6be-400b-45e6-afee-7fc443b2b23d\" (UID: \"6156b6be-400b-45e6-afee-7fc443b2b23d\") " Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.362767 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9bwx\" (UniqueName: \"kubernetes.io/projected/6156b6be-400b-45e6-afee-7fc443b2b23d-kube-api-access-h9bwx\") pod \"6156b6be-400b-45e6-afee-7fc443b2b23d\" (UID: \"6156b6be-400b-45e6-afee-7fc443b2b23d\") " Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.370846 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6156b6be-400b-45e6-afee-7fc443b2b23d-kube-api-access-h9bwx" (OuterVolumeSpecName: "kube-api-access-h9bwx") pod "6156b6be-400b-45e6-afee-7fc443b2b23d" (UID: "6156b6be-400b-45e6-afee-7fc443b2b23d"). InnerVolumeSpecName "kube-api-access-h9bwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.404566 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6156b6be-400b-45e6-afee-7fc443b2b23d-config-data" (OuterVolumeSpecName: "config-data") pod "6156b6be-400b-45e6-afee-7fc443b2b23d" (UID: "6156b6be-400b-45e6-afee-7fc443b2b23d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.407980 4744 generic.go:334] "Generic (PLEG): container finished" podID="6156b6be-400b-45e6-afee-7fc443b2b23d" containerID="d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c" exitCode=137 Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.408063 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6156b6be-400b-45e6-afee-7fc443b2b23d","Type":"ContainerDied","Data":"d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c"} Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.408133 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6156b6be-400b-45e6-afee-7fc443b2b23d","Type":"ContainerDied","Data":"89da7186eb9d228e6a6ec6eea18cf216a27b403d113fe8f6252f5700ae97b935"} Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.408159 4744 scope.go:117] "RemoveContainer" containerID="d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.408570 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.411786 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6156b6be-400b-45e6-afee-7fc443b2b23d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6156b6be-400b-45e6-afee-7fc443b2b23d" (UID: "6156b6be-400b-45e6-afee-7fc443b2b23d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.451873 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.472400 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6156b6be-400b-45e6-afee-7fc443b2b23d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.472436 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9bwx\" (UniqueName: \"kubernetes.io/projected/6156b6be-400b-45e6-afee-7fc443b2b23d-kube-api-access-h9bwx\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.472450 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6156b6be-400b-45e6-afee-7fc443b2b23d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.484262 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.485019 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.494805 4744 scope.go:117] "RemoveContainer" containerID="d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c" Oct 08 09:31:13 crc kubenswrapper[4744]: E1008 09:31:13.496813 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c\": container with ID starting with d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c not found: ID does not exist" containerID="d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.496853 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c"} err="failed to get container status \"d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c\": rpc error: code = NotFound desc = could not find container \"d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c\": container with ID starting with d6a1346a59c5ee0872f3e9b68d51ad984e26427123edb965f6699b5a3624860c not found: ID does not exist" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.751888 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.770585 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.805035 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 09:31:13 crc kubenswrapper[4744]: E1008 09:31:13.805601 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6156b6be-400b-45e6-afee-7fc443b2b23d" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.805625 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6156b6be-400b-45e6-afee-7fc443b2b23d" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.805848 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6156b6be-400b-45e6-afee-7fc443b2b23d" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.806610 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.810404 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.814328 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.830648 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.846755 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.923053 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trz4b\" (UniqueName: \"kubernetes.io/projected/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-kube-api-access-trz4b\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.923276 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.923319 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.923409 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:13 crc kubenswrapper[4744]: I1008 09:31:13.923450 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.025234 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trz4b\" (UniqueName: \"kubernetes.io/projected/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-kube-api-access-trz4b\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.025390 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.025421 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.026258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.026297 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.030715 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.032102 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.040875 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.054044 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.066027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trz4b\" (UniqueName: \"kubernetes.io/projected/a197fde0-ffb1-4ead-bdd8-acd2e60bc97f-kube-api-access-trz4b\") pod \"nova-cell1-novncproxy-0\" (UID: \"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.133652 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.428259 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 09:31:14 crc kubenswrapper[4744]: I1008 09:31:14.699081 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 09:31:15 crc kubenswrapper[4744]: I1008 09:31:15.482184 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6156b6be-400b-45e6-afee-7fc443b2b23d" path="/var/lib/kubelet/pods/6156b6be-400b-45e6-afee-7fc443b2b23d/volumes" Oct 08 09:31:15 crc kubenswrapper[4744]: I1008 09:31:15.489806 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f","Type":"ContainerStarted","Data":"3b63151b063dc565ba7bb871b6375ea2e9cb05e941850b734f08205b0d466ec7"} Oct 08 09:31:15 crc kubenswrapper[4744]: I1008 09:31:15.489849 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a197fde0-ffb1-4ead-bdd8-acd2e60bc97f","Type":"ContainerStarted","Data":"1d32d844c78d0b60514b1609d410fc8f10d599d729f75dc51727c34db24bc838"} Oct 08 09:31:15 crc kubenswrapper[4744]: I1008 09:31:15.708400 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 09:31:15 crc kubenswrapper[4744]: I1008 09:31:15.708479 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 09:31:15 crc kubenswrapper[4744]: I1008 09:31:15.709587 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 09:31:15 crc kubenswrapper[4744]: I1008 09:31:15.710594 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 09:31:15 crc kubenswrapper[4744]: I1008 09:31:15.713145 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 09:31:15 crc kubenswrapper[4744]: I1008 09:31:15.714530 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.068213 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-665946c669-qmkg4"] Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.069862 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.081425 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-ovsdbserver-nb\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.081508 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-dns-svc\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.081632 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-config\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.081662 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpsgf\" (UniqueName: \"kubernetes.io/projected/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-kube-api-access-dpsgf\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.081694 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-ovsdbserver-sb\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.099993 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-665946c669-qmkg4"] Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.184584 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-ovsdbserver-nb\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.184652 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-dns-svc\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.184872 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-config\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.184914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpsgf\" (UniqueName: \"kubernetes.io/projected/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-kube-api-access-dpsgf\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.184963 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-ovsdbserver-sb\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.185751 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-ovsdbserver-nb\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.186103 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-ovsdbserver-sb\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.186401 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-config\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.186981 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-dns-svc\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.247274 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpsgf\" (UniqueName: \"kubernetes.io/projected/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-kube-api-access-dpsgf\") pod \"dnsmasq-dns-665946c669-qmkg4\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.414360 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:16 crc kubenswrapper[4744]: I1008 09:31:16.528821 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.528795034 podStartE2EDuration="3.528795034s" podCreationTimestamp="2025-10-08 09:31:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:31:16.521942746 +0000 UTC m=+1171.769587985" watchObservedRunningTime="2025-10-08 09:31:16.528795034 +0000 UTC m=+1171.776440273" Oct 08 09:31:17 crc kubenswrapper[4744]: I1008 09:31:17.027471 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-665946c669-qmkg4"] Oct 08 09:31:17 crc kubenswrapper[4744]: I1008 09:31:17.506766 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665946c669-qmkg4" event={"ID":"3d6356c2-49f8-46b8-a5cc-6aae66240e2e","Type":"ContainerStarted","Data":"d82723004149c36ae1131c8bead029150cb577b19c90bfa662f94a86871d798e"} Oct 08 09:31:18 crc kubenswrapper[4744]: I1008 09:31:18.518630 4744 generic.go:334] "Generic (PLEG): container finished" podID="3d6356c2-49f8-46b8-a5cc-6aae66240e2e" containerID="758b3491697cca71f5668ef8a725447d4f6cd2f9a79f86722fbd89463c05259c" exitCode=0 Oct 08 09:31:18 crc kubenswrapper[4744]: I1008 09:31:18.518822 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665946c669-qmkg4" event={"ID":"3d6356c2-49f8-46b8-a5cc-6aae66240e2e","Type":"ContainerDied","Data":"758b3491697cca71f5668ef8a725447d4f6cd2f9a79f86722fbd89463c05259c"} Oct 08 09:31:19 crc kubenswrapper[4744]: I1008 09:31:19.134784 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:19 crc kubenswrapper[4744]: I1008 09:31:19.534649 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665946c669-qmkg4" event={"ID":"3d6356c2-49f8-46b8-a5cc-6aae66240e2e","Type":"ContainerStarted","Data":"b3a7a841f7e53db91817f6d3673002b5aa6c516f49d951f043223801430cbeed"} Oct 08 09:31:19 crc kubenswrapper[4744]: I1008 09:31:19.535525 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:19 crc kubenswrapper[4744]: I1008 09:31:19.560244 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:31:19 crc kubenswrapper[4744]: I1008 09:31:19.560531 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" containerName="nova-api-log" containerID="cri-o://ea33cb02115338e3fabbb430fe5159429bf7ce934db51851c504a0c1ceca3574" gracePeriod=30 Oct 08 09:31:19 crc kubenswrapper[4744]: I1008 09:31:19.560909 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" containerName="nova-api-api" containerID="cri-o://cffa31b9a562fe8c7bfb9599d87bf1723bf56195baef526033907924db285758" gracePeriod=30 Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.036074 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-665946c669-qmkg4" podStartSLOduration=4.036053093 podStartE2EDuration="4.036053093s" podCreationTimestamp="2025-10-08 09:31:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:31:19.571628802 +0000 UTC m=+1174.819274051" watchObservedRunningTime="2025-10-08 09:31:20.036053093 +0000 UTC m=+1175.283698332" Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.047899 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.048236 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="ceilometer-central-agent" containerID="cri-o://cf1131054f4311986d504bba63174293e5bcdd172e821fe26b8ccef862962f8a" gracePeriod=30 Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.048414 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="proxy-httpd" containerID="cri-o://df810bf0265cc387e02f67b03de4581a466bbd1ff5cce2b24ecb5ab373fcfdc3" gracePeriod=30 Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.048503 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="ceilometer-notification-agent" containerID="cri-o://c1695999861d43b9a189dd373e22e1ce44420c20beb3d3b721db7a2f53faed1a" gracePeriod=30 Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.048603 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="sg-core" containerID="cri-o://15145ab3881f5586c9b41f8450412a1a93c9768f6c966b73cf80d81c5dd34897" gracePeriod=30 Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.221780 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.184:3000/\": EOF" Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.545492 4744 generic.go:334] "Generic (PLEG): container finished" podID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" containerID="ea33cb02115338e3fabbb430fe5159429bf7ce934db51851c504a0c1ceca3574" exitCode=143 Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.545597 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"729a216f-b9ac-400e-bae4-c3cfe7e4023d","Type":"ContainerDied","Data":"ea33cb02115338e3fabbb430fe5159429bf7ce934db51851c504a0c1ceca3574"} Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.549461 4744 generic.go:334] "Generic (PLEG): container finished" podID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerID="df810bf0265cc387e02f67b03de4581a466bbd1ff5cce2b24ecb5ab373fcfdc3" exitCode=0 Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.549482 4744 generic.go:334] "Generic (PLEG): container finished" podID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerID="15145ab3881f5586c9b41f8450412a1a93c9768f6c966b73cf80d81c5dd34897" exitCode=2 Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.549508 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103","Type":"ContainerDied","Data":"df810bf0265cc387e02f67b03de4581a466bbd1ff5cce2b24ecb5ab373fcfdc3"} Oct 08 09:31:20 crc kubenswrapper[4744]: I1008 09:31:20.549555 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103","Type":"ContainerDied","Data":"15145ab3881f5586c9b41f8450412a1a93c9768f6c966b73cf80d81c5dd34897"} Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.563693 4744 generic.go:334] "Generic (PLEG): container finished" podID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerID="c1695999861d43b9a189dd373e22e1ce44420c20beb3d3b721db7a2f53faed1a" exitCode=0 Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.564062 4744 generic.go:334] "Generic (PLEG): container finished" podID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerID="cf1131054f4311986d504bba63174293e5bcdd172e821fe26b8ccef862962f8a" exitCode=0 Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.563726 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103","Type":"ContainerDied","Data":"c1695999861d43b9a189dd373e22e1ce44420c20beb3d3b721db7a2f53faed1a"} Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.564130 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103","Type":"ContainerDied","Data":"cf1131054f4311986d504bba63174293e5bcdd172e821fe26b8ccef862962f8a"} Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.678271 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.815460 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-run-httpd\") pod \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.815564 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-log-httpd\") pod \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.815635 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-ceilometer-tls-certs\") pod \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.815717 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-sg-core-conf-yaml\") pod \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.815791 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-scripts\") pod \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.815902 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rznf8\" (UniqueName: \"kubernetes.io/projected/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-kube-api-access-rznf8\") pod \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.815948 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-combined-ca-bundle\") pod \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.815972 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-config-data\") pod \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\" (UID: \"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103\") " Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.816180 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" (UID: "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.816822 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.817229 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" (UID: "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.838690 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-kube-api-access-rznf8" (OuterVolumeSpecName: "kube-api-access-rznf8") pod "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" (UID: "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103"). InnerVolumeSpecName "kube-api-access-rznf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.839072 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-scripts" (OuterVolumeSpecName: "scripts") pod "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" (UID: "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.876694 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" (UID: "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.893623 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" (UID: "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.915347 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" (UID: "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.918354 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rznf8\" (UniqueName: \"kubernetes.io/projected/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-kube-api-access-rznf8\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.918388 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.918406 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.918416 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.918424 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.918434 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:21 crc kubenswrapper[4744]: I1008 09:31:21.962640 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-config-data" (OuterVolumeSpecName: "config-data") pod "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" (UID: "188a0d94-bf8c-44cc-bbe2-8cd1d0e75103"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.020988 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.578114 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"188a0d94-bf8c-44cc-bbe2-8cd1d0e75103","Type":"ContainerDied","Data":"f62efb4200d9c821b858b7a296271a5c232a3099e87a6fcc3838d716d5c323d3"} Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.578214 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.578611 4744 scope.go:117] "RemoveContainer" containerID="df810bf0265cc387e02f67b03de4581a466bbd1ff5cce2b24ecb5ab373fcfdc3" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.613259 4744 scope.go:117] "RemoveContainer" containerID="15145ab3881f5586c9b41f8450412a1a93c9768f6c966b73cf80d81c5dd34897" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.621885 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.633115 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.641456 4744 scope.go:117] "RemoveContainer" containerID="c1695999861d43b9a189dd373e22e1ce44420c20beb3d3b721db7a2f53faed1a" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.672392 4744 scope.go:117] "RemoveContainer" containerID="cf1131054f4311986d504bba63174293e5bcdd172e821fe26b8ccef862962f8a" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.682866 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:31:22 crc kubenswrapper[4744]: E1008 09:31:22.683298 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="ceilometer-central-agent" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.683318 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="ceilometer-central-agent" Oct 08 09:31:22 crc kubenswrapper[4744]: E1008 09:31:22.683339 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="sg-core" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.683346 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="sg-core" Oct 08 09:31:22 crc kubenswrapper[4744]: E1008 09:31:22.683353 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="proxy-httpd" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.683360 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="proxy-httpd" Oct 08 09:31:22 crc kubenswrapper[4744]: E1008 09:31:22.683391 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="ceilometer-notification-agent" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.683397 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="ceilometer-notification-agent" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.683701 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="sg-core" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.683715 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="proxy-httpd" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.683818 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="ceilometer-central-agent" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.683835 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" containerName="ceilometer-notification-agent" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.685603 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.696505 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.696804 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.696964 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.708378 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.736053 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6a8bac2-8a56-45d4-a354-7814253f473d-log-httpd\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.736156 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-config-data\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.736173 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.736231 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-scripts\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.736258 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6a8bac2-8a56-45d4-a354-7814253f473d-run-httpd\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.736421 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjxq2\" (UniqueName: \"kubernetes.io/projected/a6a8bac2-8a56-45d4-a354-7814253f473d-kube-api-access-gjxq2\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.736456 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.736519 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.838529 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6a8bac2-8a56-45d4-a354-7814253f473d-log-httpd\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.838593 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-config-data\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.838617 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.838647 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-scripts\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.838678 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6a8bac2-8a56-45d4-a354-7814253f473d-run-httpd\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.838762 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjxq2\" (UniqueName: \"kubernetes.io/projected/a6a8bac2-8a56-45d4-a354-7814253f473d-kube-api-access-gjxq2\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.838796 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.838828 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.839700 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6a8bac2-8a56-45d4-a354-7814253f473d-log-httpd\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.840567 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6a8bac2-8a56-45d4-a354-7814253f473d-run-httpd\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.850098 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-config-data\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.850809 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.851225 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-scripts\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.852678 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.854129 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:22 crc kubenswrapper[4744]: I1008 09:31:22.863969 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjxq2\" (UniqueName: \"kubernetes.io/projected/a6a8bac2-8a56-45d4-a354-7814253f473d-kube-api-access-gjxq2\") pod \"ceilometer-0\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " pod="openstack/ceilometer-0" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.011969 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.469824 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="188a0d94-bf8c-44cc-bbe2-8cd1d0e75103" path="/var/lib/kubelet/pods/188a0d94-bf8c-44cc-bbe2-8cd1d0e75103/volumes" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.607237 4744 generic.go:334] "Generic (PLEG): container finished" podID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" containerID="cffa31b9a562fe8c7bfb9599d87bf1723bf56195baef526033907924db285758" exitCode=0 Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.607661 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"729a216f-b9ac-400e-bae4-c3cfe7e4023d","Type":"ContainerDied","Data":"cffa31b9a562fe8c7bfb9599d87bf1723bf56195baef526033907924db285758"} Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.607715 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"729a216f-b9ac-400e-bae4-c3cfe7e4023d","Type":"ContainerDied","Data":"cecf06170045e24821547bfb8c4db488244085a09e7dd80427ad76ec2e211965"} Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.607730 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cecf06170045e24821547bfb8c4db488244085a09e7dd80427ad76ec2e211965" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.609392 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.661380 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729a216f-b9ac-400e-bae4-c3cfe7e4023d-config-data\") pod \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.661440 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729a216f-b9ac-400e-bae4-c3cfe7e4023d-combined-ca-bundle\") pod \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.661565 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/729a216f-b9ac-400e-bae4-c3cfe7e4023d-logs\") pod \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.661626 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxxms\" (UniqueName: \"kubernetes.io/projected/729a216f-b9ac-400e-bae4-c3cfe7e4023d-kube-api-access-nxxms\") pod \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\" (UID: \"729a216f-b9ac-400e-bae4-c3cfe7e4023d\") " Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.662707 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/729a216f-b9ac-400e-bae4-c3cfe7e4023d-logs" (OuterVolumeSpecName: "logs") pod "729a216f-b9ac-400e-bae4-c3cfe7e4023d" (UID: "729a216f-b9ac-400e-bae4-c3cfe7e4023d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.689204 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/729a216f-b9ac-400e-bae4-c3cfe7e4023d-kube-api-access-nxxms" (OuterVolumeSpecName: "kube-api-access-nxxms") pod "729a216f-b9ac-400e-bae4-c3cfe7e4023d" (UID: "729a216f-b9ac-400e-bae4-c3cfe7e4023d"). InnerVolumeSpecName "kube-api-access-nxxms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.725587 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/729a216f-b9ac-400e-bae4-c3cfe7e4023d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "729a216f-b9ac-400e-bae4-c3cfe7e4023d" (UID: "729a216f-b9ac-400e-bae4-c3cfe7e4023d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.738883 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/729a216f-b9ac-400e-bae4-c3cfe7e4023d-config-data" (OuterVolumeSpecName: "config-data") pod "729a216f-b9ac-400e-bae4-c3cfe7e4023d" (UID: "729a216f-b9ac-400e-bae4-c3cfe7e4023d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.763559 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/729a216f-b9ac-400e-bae4-c3cfe7e4023d-logs\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.763600 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxxms\" (UniqueName: \"kubernetes.io/projected/729a216f-b9ac-400e-bae4-c3cfe7e4023d-kube-api-access-nxxms\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.763614 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729a216f-b9ac-400e-bae4-c3cfe7e4023d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.763624 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729a216f-b9ac-400e-bae4-c3cfe7e4023d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.787328 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 09:31:23 crc kubenswrapper[4744]: I1008 09:31:23.792759 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.134323 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.155255 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.634350 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.634338 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6a8bac2-8a56-45d4-a354-7814253f473d","Type":"ContainerStarted","Data":"dcfa0834c1f78532d0765c6898522eb0b5259139bd7ec1b122acddae4f284561"} Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.636585 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6a8bac2-8a56-45d4-a354-7814253f473d","Type":"ContainerStarted","Data":"9184172c78395e43513cdcde59ed391780fc131986ce75aaeb71e4c7cf33ad0b"} Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.669348 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.688436 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.697448 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.720359 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 09:31:24 crc kubenswrapper[4744]: E1008 09:31:24.720951 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" containerName="nova-api-api" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.720973 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" containerName="nova-api-api" Oct 08 09:31:24 crc kubenswrapper[4744]: E1008 09:31:24.721019 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" containerName="nova-api-log" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.721029 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" containerName="nova-api-log" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.721223 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" containerName="nova-api-log" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.721239 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" containerName="nova-api-api" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.722650 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.729315 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.729682 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.729796 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.755451 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.789680 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-public-tls-certs\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.789875 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-config-data\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.789907 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.789990 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdb30ed4-01f7-4555-ac98-0ad496f7af40-logs\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.790159 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvr9l\" (UniqueName: \"kubernetes.io/projected/bdb30ed4-01f7-4555-ac98-0ad496f7af40-kube-api-access-pvr9l\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.790218 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.893563 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-config-data\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.893622 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.893663 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdb30ed4-01f7-4555-ac98-0ad496f7af40-logs\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.893723 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvr9l\" (UniqueName: \"kubernetes.io/projected/bdb30ed4-01f7-4555-ac98-0ad496f7af40-kube-api-access-pvr9l\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.893750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.893801 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-public-tls-certs\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.895622 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdb30ed4-01f7-4555-ac98-0ad496f7af40-logs\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.905349 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-public-tls-certs\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.908057 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-internal-tls-certs\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.916258 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.916813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-config-data\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.921238 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvr9l\" (UniqueName: \"kubernetes.io/projected/bdb30ed4-01f7-4555-ac98-0ad496f7af40-kube-api-access-pvr9l\") pod \"nova-api-0\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " pod="openstack/nova-api-0" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.986648 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-62khr"] Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.988087 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.993302 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 08 09:31:24 crc kubenswrapper[4744]: I1008 09:31:24.993721 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.016280 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-62khr"] Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.065844 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.114025 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-scripts\") pod \"nova-cell1-cell-mapping-62khr\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.114075 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-config-data\") pod \"nova-cell1-cell-mapping-62khr\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.114250 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l29dc\" (UniqueName: \"kubernetes.io/projected/fc006ede-60d5-417e-bde3-e3bca31b912b-kube-api-access-l29dc\") pod \"nova-cell1-cell-mapping-62khr\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.114414 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-62khr\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.220678 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-scripts\") pod \"nova-cell1-cell-mapping-62khr\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.221017 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-config-data\") pod \"nova-cell1-cell-mapping-62khr\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.221165 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l29dc\" (UniqueName: \"kubernetes.io/projected/fc006ede-60d5-417e-bde3-e3bca31b912b-kube-api-access-l29dc\") pod \"nova-cell1-cell-mapping-62khr\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.221236 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-62khr\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.230522 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-62khr\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.231289 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-config-data\") pod \"nova-cell1-cell-mapping-62khr\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.235249 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-scripts\") pod \"nova-cell1-cell-mapping-62khr\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.281086 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l29dc\" (UniqueName: \"kubernetes.io/projected/fc006ede-60d5-417e-bde3-e3bca31b912b-kube-api-access-l29dc\") pod \"nova-cell1-cell-mapping-62khr\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.356624 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.486469 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="729a216f-b9ac-400e-bae4-c3cfe7e4023d" path="/var/lib/kubelet/pods/729a216f-b9ac-400e-bae4-c3cfe7e4023d/volumes" Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.634883 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.665966 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6a8bac2-8a56-45d4-a354-7814253f473d","Type":"ContainerStarted","Data":"73753fe1ac100ccb50f054243e0ae9aca6a241a45c8569445513a0fbf6c18f32"} Oct 08 09:31:25 crc kubenswrapper[4744]: I1008 09:31:25.911153 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-62khr"] Oct 08 09:31:26 crc kubenswrapper[4744]: I1008 09:31:26.417624 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:31:26 crc kubenswrapper[4744]: I1008 09:31:26.565073 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75fb48c489-dbtdd"] Oct 08 09:31:26 crc kubenswrapper[4744]: I1008 09:31:26.585271 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" podUID="a81b3a01-b0c5-4a5f-b50e-6c28837ff977" containerName="dnsmasq-dns" containerID="cri-o://8f697f3d0acca97b854bba900c82fab6e7dee4be65cc9e848d46bf9382b30c00" gracePeriod=10 Oct 08 09:31:26 crc kubenswrapper[4744]: I1008 09:31:26.713195 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bdb30ed4-01f7-4555-ac98-0ad496f7af40","Type":"ContainerStarted","Data":"47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be"} Oct 08 09:31:26 crc kubenswrapper[4744]: I1008 09:31:26.713257 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bdb30ed4-01f7-4555-ac98-0ad496f7af40","Type":"ContainerStarted","Data":"7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0"} Oct 08 09:31:26 crc kubenswrapper[4744]: I1008 09:31:26.713269 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bdb30ed4-01f7-4555-ac98-0ad496f7af40","Type":"ContainerStarted","Data":"701827a267963e69cedd229954cc7c0a1b94cf69e1e7112c7166bfb1256c147a"} Oct 08 09:31:26 crc kubenswrapper[4744]: I1008 09:31:26.716406 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-62khr" event={"ID":"fc006ede-60d5-417e-bde3-e3bca31b912b","Type":"ContainerStarted","Data":"519f31b51a00c71469332f0ce2b05612c09ca86e54492eacd13c5767b8c460d8"} Oct 08 09:31:26 crc kubenswrapper[4744]: I1008 09:31:26.716464 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-62khr" event={"ID":"fc006ede-60d5-417e-bde3-e3bca31b912b","Type":"ContainerStarted","Data":"a0a760f6cd6ee84785f0ebdac5816bc66a861aba4dc334dde617a52b1d2de10c"} Oct 08 09:31:26 crc kubenswrapper[4744]: I1008 09:31:26.762986 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6a8bac2-8a56-45d4-a354-7814253f473d","Type":"ContainerStarted","Data":"291f257a2849b68458000e9bd8db98c8043e1f8bd5f6ebc02775d9dde2f6caf9"} Oct 08 09:31:26 crc kubenswrapper[4744]: I1008 09:31:26.788511 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.788486391 podStartE2EDuration="2.788486391s" podCreationTimestamp="2025-10-08 09:31:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:31:26.748871839 +0000 UTC m=+1181.996517078" watchObservedRunningTime="2025-10-08 09:31:26.788486391 +0000 UTC m=+1182.036131630" Oct 08 09:31:26 crc kubenswrapper[4744]: I1008 09:31:26.819512 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-62khr" podStartSLOduration=2.819487676 podStartE2EDuration="2.819487676s" podCreationTimestamp="2025-10-08 09:31:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:31:26.801920551 +0000 UTC m=+1182.049565790" watchObservedRunningTime="2025-10-08 09:31:26.819487676 +0000 UTC m=+1182.067132915" Oct 08 09:31:27 crc kubenswrapper[4744]: I1008 09:31:27.194298 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" podUID="a81b3a01-b0c5-4a5f-b50e-6c28837ff977" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.175:5353: connect: connection refused" Oct 08 09:31:27 crc kubenswrapper[4744]: I1008 09:31:27.780253 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6a8bac2-8a56-45d4-a354-7814253f473d","Type":"ContainerStarted","Data":"ee689b75f227cdea446701892d4c60fef6c78f2ae31fae82c065c457d06e1169"} Oct 08 09:31:27 crc kubenswrapper[4744]: I1008 09:31:27.782279 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 09:31:27 crc kubenswrapper[4744]: I1008 09:31:27.786022 4744 generic.go:334] "Generic (PLEG): container finished" podID="a81b3a01-b0c5-4a5f-b50e-6c28837ff977" containerID="8f697f3d0acca97b854bba900c82fab6e7dee4be65cc9e848d46bf9382b30c00" exitCode=0 Oct 08 09:31:27 crc kubenswrapper[4744]: I1008 09:31:27.786194 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" event={"ID":"a81b3a01-b0c5-4a5f-b50e-6c28837ff977","Type":"ContainerDied","Data":"8f697f3d0acca97b854bba900c82fab6e7dee4be65cc9e848d46bf9382b30c00"} Oct 08 09:31:27 crc kubenswrapper[4744]: I1008 09:31:27.822966 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.583194917 podStartE2EDuration="5.822945607s" podCreationTimestamp="2025-10-08 09:31:22 +0000 UTC" firstStartedPulling="2025-10-08 09:31:23.787111897 +0000 UTC m=+1179.034757136" lastFinishedPulling="2025-10-08 09:31:27.026862597 +0000 UTC m=+1182.274507826" observedRunningTime="2025-10-08 09:31:27.82126414 +0000 UTC m=+1183.068909379" watchObservedRunningTime="2025-10-08 09:31:27.822945607 +0000 UTC m=+1183.070590846" Oct 08 09:31:27 crc kubenswrapper[4744]: I1008 09:31:27.895041 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.020255 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-ovsdbserver-sb\") pod \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.020687 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcjkg\" (UniqueName: \"kubernetes.io/projected/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-kube-api-access-dcjkg\") pod \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.020835 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-dns-svc\") pod \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.020928 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-ovsdbserver-nb\") pod \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.021008 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-config\") pod \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\" (UID: \"a81b3a01-b0c5-4a5f-b50e-6c28837ff977\") " Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.048686 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-kube-api-access-dcjkg" (OuterVolumeSpecName: "kube-api-access-dcjkg") pod "a81b3a01-b0c5-4a5f-b50e-6c28837ff977" (UID: "a81b3a01-b0c5-4a5f-b50e-6c28837ff977"). InnerVolumeSpecName "kube-api-access-dcjkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.088316 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-config" (OuterVolumeSpecName: "config") pod "a81b3a01-b0c5-4a5f-b50e-6c28837ff977" (UID: "a81b3a01-b0c5-4a5f-b50e-6c28837ff977"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.106059 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a81b3a01-b0c5-4a5f-b50e-6c28837ff977" (UID: "a81b3a01-b0c5-4a5f-b50e-6c28837ff977"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.113809 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a81b3a01-b0c5-4a5f-b50e-6c28837ff977" (UID: "a81b3a01-b0c5-4a5f-b50e-6c28837ff977"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.124192 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.124229 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.124238 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.124249 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcjkg\" (UniqueName: \"kubernetes.io/projected/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-kube-api-access-dcjkg\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.135009 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a81b3a01-b0c5-4a5f-b50e-6c28837ff977" (UID: "a81b3a01-b0c5-4a5f-b50e-6c28837ff977"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.226890 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a81b3a01-b0c5-4a5f-b50e-6c28837ff977-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.799713 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" event={"ID":"a81b3a01-b0c5-4a5f-b50e-6c28837ff977","Type":"ContainerDied","Data":"024b2df5a5d0e337de61d42f8330eb105bff527548ae5deab74b3d1a1c898b4e"} Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.799794 4744 scope.go:117] "RemoveContainer" containerID="8f697f3d0acca97b854bba900c82fab6e7dee4be65cc9e848d46bf9382b30c00" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.799748 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75fb48c489-dbtdd" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.839669 4744 scope.go:117] "RemoveContainer" containerID="f76d20950f06678527ff0d329d8b1b63bd12a28221fe466d63e6c07c89bc3370" Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.855752 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75fb48c489-dbtdd"] Oct 08 09:31:28 crc kubenswrapper[4744]: I1008 09:31:28.866609 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75fb48c489-dbtdd"] Oct 08 09:31:29 crc kubenswrapper[4744]: I1008 09:31:29.466162 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a81b3a01-b0c5-4a5f-b50e-6c28837ff977" path="/var/lib/kubelet/pods/a81b3a01-b0c5-4a5f-b50e-6c28837ff977/volumes" Oct 08 09:31:32 crc kubenswrapper[4744]: I1008 09:31:32.847383 4744 generic.go:334] "Generic (PLEG): container finished" podID="fc006ede-60d5-417e-bde3-e3bca31b912b" containerID="519f31b51a00c71469332f0ce2b05612c09ca86e54492eacd13c5767b8c460d8" exitCode=0 Oct 08 09:31:32 crc kubenswrapper[4744]: I1008 09:31:32.847610 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-62khr" event={"ID":"fc006ede-60d5-417e-bde3-e3bca31b912b","Type":"ContainerDied","Data":"519f31b51a00c71469332f0ce2b05612c09ca86e54492eacd13c5767b8c460d8"} Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.224405 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.263360 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-combined-ca-bundle\") pod \"fc006ede-60d5-417e-bde3-e3bca31b912b\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.263497 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l29dc\" (UniqueName: \"kubernetes.io/projected/fc006ede-60d5-417e-bde3-e3bca31b912b-kube-api-access-l29dc\") pod \"fc006ede-60d5-417e-bde3-e3bca31b912b\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.263566 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-scripts\") pod \"fc006ede-60d5-417e-bde3-e3bca31b912b\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.263609 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-config-data\") pod \"fc006ede-60d5-417e-bde3-e3bca31b912b\" (UID: \"fc006ede-60d5-417e-bde3-e3bca31b912b\") " Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.274849 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-scripts" (OuterVolumeSpecName: "scripts") pod "fc006ede-60d5-417e-bde3-e3bca31b912b" (UID: "fc006ede-60d5-417e-bde3-e3bca31b912b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.279205 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc006ede-60d5-417e-bde3-e3bca31b912b-kube-api-access-l29dc" (OuterVolumeSpecName: "kube-api-access-l29dc") pod "fc006ede-60d5-417e-bde3-e3bca31b912b" (UID: "fc006ede-60d5-417e-bde3-e3bca31b912b"). InnerVolumeSpecName "kube-api-access-l29dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.295267 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-config-data" (OuterVolumeSpecName: "config-data") pod "fc006ede-60d5-417e-bde3-e3bca31b912b" (UID: "fc006ede-60d5-417e-bde3-e3bca31b912b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.303525 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc006ede-60d5-417e-bde3-e3bca31b912b" (UID: "fc006ede-60d5-417e-bde3-e3bca31b912b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.368630 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.368766 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l29dc\" (UniqueName: \"kubernetes.io/projected/fc006ede-60d5-417e-bde3-e3bca31b912b-kube-api-access-l29dc\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.368806 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.368832 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc006ede-60d5-417e-bde3-e3bca31b912b-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.867633 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-62khr" event={"ID":"fc006ede-60d5-417e-bde3-e3bca31b912b","Type":"ContainerDied","Data":"a0a760f6cd6ee84785f0ebdac5816bc66a861aba4dc334dde617a52b1d2de10c"} Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.867680 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0a760f6cd6ee84785f0ebdac5816bc66a861aba4dc334dde617a52b1d2de10c" Oct 08 09:31:34 crc kubenswrapper[4744]: I1008 09:31:34.867761 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-62khr" Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.067465 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.067564 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.067589 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.069290 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4e3d9501-0500-4fea-a909-0d3845a1845d" containerName="nova-scheduler-scheduler" containerID="cri-o://c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f" gracePeriod=30 Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.083042 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.127249 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.127557 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerName="nova-metadata-log" containerID="cri-o://0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21" gracePeriod=30 Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.127737 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerName="nova-metadata-metadata" containerID="cri-o://7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8" gracePeriod=30 Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.878053 4744 generic.go:334] "Generic (PLEG): container finished" podID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerID="0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21" exitCode=143 Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.878609 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" containerName="nova-api-log" containerID="cri-o://7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0" gracePeriod=30 Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.878696 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"237f9dff-1f92-46fa-a8d3-23f447f334be","Type":"ContainerDied","Data":"0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21"} Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.879133 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" containerName="nova-api-api" containerID="cri-o://47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be" gracePeriod=30 Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.882624 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.188:8774/\": EOF" Oct 08 09:31:35 crc kubenswrapper[4744]: I1008 09:31:35.882617 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.188:8774/\": EOF" Oct 08 09:31:35 crc kubenswrapper[4744]: E1008 09:31:35.905905 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 09:31:35 crc kubenswrapper[4744]: E1008 09:31:35.910548 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 09:31:35 crc kubenswrapper[4744]: E1008 09:31:35.912283 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 09:31:35 crc kubenswrapper[4744]: E1008 09:31:35.912443 4744 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4e3d9501-0500-4fea-a909-0d3845a1845d" containerName="nova-scheduler-scheduler" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.659642 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.720075 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8652l\" (UniqueName: \"kubernetes.io/projected/4e3d9501-0500-4fea-a909-0d3845a1845d-kube-api-access-8652l\") pod \"4e3d9501-0500-4fea-a909-0d3845a1845d\" (UID: \"4e3d9501-0500-4fea-a909-0d3845a1845d\") " Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.720337 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e3d9501-0500-4fea-a909-0d3845a1845d-config-data\") pod \"4e3d9501-0500-4fea-a909-0d3845a1845d\" (UID: \"4e3d9501-0500-4fea-a909-0d3845a1845d\") " Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.720488 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e3d9501-0500-4fea-a909-0d3845a1845d-combined-ca-bundle\") pod \"4e3d9501-0500-4fea-a909-0d3845a1845d\" (UID: \"4e3d9501-0500-4fea-a909-0d3845a1845d\") " Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.728616 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e3d9501-0500-4fea-a909-0d3845a1845d-kube-api-access-8652l" (OuterVolumeSpecName: "kube-api-access-8652l") pod "4e3d9501-0500-4fea-a909-0d3845a1845d" (UID: "4e3d9501-0500-4fea-a909-0d3845a1845d"). InnerVolumeSpecName "kube-api-access-8652l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.762257 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e3d9501-0500-4fea-a909-0d3845a1845d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e3d9501-0500-4fea-a909-0d3845a1845d" (UID: "4e3d9501-0500-4fea-a909-0d3845a1845d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.763529 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e3d9501-0500-4fea-a909-0d3845a1845d-config-data" (OuterVolumeSpecName: "config-data") pod "4e3d9501-0500-4fea-a909-0d3845a1845d" (UID: "4e3d9501-0500-4fea-a909-0d3845a1845d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.823054 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8652l\" (UniqueName: \"kubernetes.io/projected/4e3d9501-0500-4fea-a909-0d3845a1845d-kube-api-access-8652l\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.823094 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e3d9501-0500-4fea-a909-0d3845a1845d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.823108 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e3d9501-0500-4fea-a909-0d3845a1845d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.888789 4744 generic.go:334] "Generic (PLEG): container finished" podID="4e3d9501-0500-4fea-a909-0d3845a1845d" containerID="c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f" exitCode=0 Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.888915 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4e3d9501-0500-4fea-a909-0d3845a1845d","Type":"ContainerDied","Data":"c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f"} Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.888962 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4e3d9501-0500-4fea-a909-0d3845a1845d","Type":"ContainerDied","Data":"c34d179eb17554321ae770727d1a04eb5809363bd8d29d43d9ea7f07e84731e8"} Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.888984 4744 scope.go:117] "RemoveContainer" containerID="c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.889970 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.891161 4744 generic.go:334] "Generic (PLEG): container finished" podID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" containerID="7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0" exitCode=143 Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.891198 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bdb30ed4-01f7-4555-ac98-0ad496f7af40","Type":"ContainerDied","Data":"7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0"} Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.930965 4744 scope.go:117] "RemoveContainer" containerID="c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f" Oct 08 09:31:36 crc kubenswrapper[4744]: E1008 09:31:36.931318 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f\": container with ID starting with c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f not found: ID does not exist" containerID="c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.931347 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f"} err="failed to get container status \"c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f\": rpc error: code = NotFound desc = could not find container \"c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f\": container with ID starting with c8744d1bd8ccbf8a39232a893308b4dc2c3f1f1cc7ea55d27ee9528a303b204f not found: ID does not exist" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.944353 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.959728 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.971082 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:31:36 crc kubenswrapper[4744]: E1008 09:31:36.971700 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a81b3a01-b0c5-4a5f-b50e-6c28837ff977" containerName="dnsmasq-dns" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.971721 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a81b3a01-b0c5-4a5f-b50e-6c28837ff977" containerName="dnsmasq-dns" Oct 08 09:31:36 crc kubenswrapper[4744]: E1008 09:31:36.971760 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e3d9501-0500-4fea-a909-0d3845a1845d" containerName="nova-scheduler-scheduler" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.971770 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e3d9501-0500-4fea-a909-0d3845a1845d" containerName="nova-scheduler-scheduler" Oct 08 09:31:36 crc kubenswrapper[4744]: E1008 09:31:36.971784 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a81b3a01-b0c5-4a5f-b50e-6c28837ff977" containerName="init" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.971793 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a81b3a01-b0c5-4a5f-b50e-6c28837ff977" containerName="init" Oct 08 09:31:36 crc kubenswrapper[4744]: E1008 09:31:36.971809 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc006ede-60d5-417e-bde3-e3bca31b912b" containerName="nova-manage" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.971817 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc006ede-60d5-417e-bde3-e3bca31b912b" containerName="nova-manage" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.972040 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e3d9501-0500-4fea-a909-0d3845a1845d" containerName="nova-scheduler-scheduler" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.972069 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a81b3a01-b0c5-4a5f-b50e-6c28837ff977" containerName="dnsmasq-dns" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.972085 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc006ede-60d5-417e-bde3-e3bca31b912b" containerName="nova-manage" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.973024 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.977571 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:31:36 crc kubenswrapper[4744]: I1008 09:31:36.978961 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.027028 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6ac453-8ad5-4c02-adf0-ff88a946dd15-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ee6ac453-8ad5-4c02-adf0-ff88a946dd15\") " pod="openstack/nova-scheduler-0" Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.027140 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht264\" (UniqueName: \"kubernetes.io/projected/ee6ac453-8ad5-4c02-adf0-ff88a946dd15-kube-api-access-ht264\") pod \"nova-scheduler-0\" (UID: \"ee6ac453-8ad5-4c02-adf0-ff88a946dd15\") " pod="openstack/nova-scheduler-0" Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.027173 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6ac453-8ad5-4c02-adf0-ff88a946dd15-config-data\") pod \"nova-scheduler-0\" (UID: \"ee6ac453-8ad5-4c02-adf0-ff88a946dd15\") " pod="openstack/nova-scheduler-0" Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.128681 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6ac453-8ad5-4c02-adf0-ff88a946dd15-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ee6ac453-8ad5-4c02-adf0-ff88a946dd15\") " pod="openstack/nova-scheduler-0" Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.128856 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht264\" (UniqueName: \"kubernetes.io/projected/ee6ac453-8ad5-4c02-adf0-ff88a946dd15-kube-api-access-ht264\") pod \"nova-scheduler-0\" (UID: \"ee6ac453-8ad5-4c02-adf0-ff88a946dd15\") " pod="openstack/nova-scheduler-0" Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.130467 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6ac453-8ad5-4c02-adf0-ff88a946dd15-config-data\") pod \"nova-scheduler-0\" (UID: \"ee6ac453-8ad5-4c02-adf0-ff88a946dd15\") " pod="openstack/nova-scheduler-0" Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.133581 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee6ac453-8ad5-4c02-adf0-ff88a946dd15-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ee6ac453-8ad5-4c02-adf0-ff88a946dd15\") " pod="openstack/nova-scheduler-0" Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.135172 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee6ac453-8ad5-4c02-adf0-ff88a946dd15-config-data\") pod \"nova-scheduler-0\" (UID: \"ee6ac453-8ad5-4c02-adf0-ff88a946dd15\") " pod="openstack/nova-scheduler-0" Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.148279 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht264\" (UniqueName: \"kubernetes.io/projected/ee6ac453-8ad5-4c02-adf0-ff88a946dd15-kube-api-access-ht264\") pod \"nova-scheduler-0\" (UID: \"ee6ac453-8ad5-4c02-adf0-ff88a946dd15\") " pod="openstack/nova-scheduler-0" Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.292463 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.511231 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e3d9501-0500-4fea-a909-0d3845a1845d" path="/var/lib/kubelet/pods/4e3d9501-0500-4fea-a909-0d3845a1845d/volumes" Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.823917 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 09:31:37 crc kubenswrapper[4744]: I1008 09:31:37.916938 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ee6ac453-8ad5-4c02-adf0-ff88a946dd15","Type":"ContainerStarted","Data":"0565dd111522bbd07636d5c1972a848fafaaf7be5c60cb06d745e4d9b16f4f4c"} Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.437031 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.180:8775/\": dial tcp 10.217.0.180:8775: connect: connection refused" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.437096 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.180:8775/\": dial tcp 10.217.0.180:8775: connect: connection refused" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.741823 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.866963 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7f5d\" (UniqueName: \"kubernetes.io/projected/237f9dff-1f92-46fa-a8d3-23f447f334be-kube-api-access-g7f5d\") pod \"237f9dff-1f92-46fa-a8d3-23f447f334be\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.867072 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-nova-metadata-tls-certs\") pod \"237f9dff-1f92-46fa-a8d3-23f447f334be\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.867131 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-config-data\") pod \"237f9dff-1f92-46fa-a8d3-23f447f334be\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.867222 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-combined-ca-bundle\") pod \"237f9dff-1f92-46fa-a8d3-23f447f334be\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.867341 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/237f9dff-1f92-46fa-a8d3-23f447f334be-logs\") pod \"237f9dff-1f92-46fa-a8d3-23f447f334be\" (UID: \"237f9dff-1f92-46fa-a8d3-23f447f334be\") " Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.868455 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/237f9dff-1f92-46fa-a8d3-23f447f334be-logs" (OuterVolumeSpecName: "logs") pod "237f9dff-1f92-46fa-a8d3-23f447f334be" (UID: "237f9dff-1f92-46fa-a8d3-23f447f334be"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.872563 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/237f9dff-1f92-46fa-a8d3-23f447f334be-kube-api-access-g7f5d" (OuterVolumeSpecName: "kube-api-access-g7f5d") pod "237f9dff-1f92-46fa-a8d3-23f447f334be" (UID: "237f9dff-1f92-46fa-a8d3-23f447f334be"). InnerVolumeSpecName "kube-api-access-g7f5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.902167 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-config-data" (OuterVolumeSpecName: "config-data") pod "237f9dff-1f92-46fa-a8d3-23f447f334be" (UID: "237f9dff-1f92-46fa-a8d3-23f447f334be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.919307 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "237f9dff-1f92-46fa-a8d3-23f447f334be" (UID: "237f9dff-1f92-46fa-a8d3-23f447f334be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.924010 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "237f9dff-1f92-46fa-a8d3-23f447f334be" (UID: "237f9dff-1f92-46fa-a8d3-23f447f334be"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.937056 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ee6ac453-8ad5-4c02-adf0-ff88a946dd15","Type":"ContainerStarted","Data":"2f19d600b637c7551a871fb404a82fcfcbc4979842f6e93016e662b204634306"} Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.941557 4744 generic.go:334] "Generic (PLEG): container finished" podID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerID="7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8" exitCode=0 Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.941603 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"237f9dff-1f92-46fa-a8d3-23f447f334be","Type":"ContainerDied","Data":"7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8"} Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.941633 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"237f9dff-1f92-46fa-a8d3-23f447f334be","Type":"ContainerDied","Data":"2cebdb2c9a7405308921569794db4071a1ed0e42d8befef5863153233504dab5"} Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.941651 4744 scope.go:117] "RemoveContainer" containerID="7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.941676 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.962444 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.9624257419999998 podStartE2EDuration="2.962425742s" podCreationTimestamp="2025-10-08 09:31:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:31:38.959132542 +0000 UTC m=+1194.206777801" watchObservedRunningTime="2025-10-08 09:31:38.962425742 +0000 UTC m=+1194.210070981" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.970312 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.970875 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.970889 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/237f9dff-1f92-46fa-a8d3-23f447f334be-logs\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.970899 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7f5d\" (UniqueName: \"kubernetes.io/projected/237f9dff-1f92-46fa-a8d3-23f447f334be-kube-api-access-g7f5d\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.970910 4744 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/237f9dff-1f92-46fa-a8d3-23f447f334be-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.976588 4744 scope.go:117] "RemoveContainer" containerID="0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21" Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.983742 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:31:38 crc kubenswrapper[4744]: I1008 09:31:38.999025 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.002893 4744 scope.go:117] "RemoveContainer" containerID="7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8" Oct 08 09:31:39 crc kubenswrapper[4744]: E1008 09:31:39.003638 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8\": container with ID starting with 7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8 not found: ID does not exist" containerID="7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.003667 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8"} err="failed to get container status \"7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8\": rpc error: code = NotFound desc = could not find container \"7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8\": container with ID starting with 7f4fe0a0c5f1e69fe28d05a805849413b6fd71cc97baa6ee5d977761f6c457a8 not found: ID does not exist" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.003710 4744 scope.go:117] "RemoveContainer" containerID="0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21" Oct 08 09:31:39 crc kubenswrapper[4744]: E1008 09:31:39.004122 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21\": container with ID starting with 0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21 not found: ID does not exist" containerID="0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.004143 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21"} err="failed to get container status \"0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21\": rpc error: code = NotFound desc = could not find container \"0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21\": container with ID starting with 0ad57bcc3ac5136e568cadd0e5d83d6ab56c6b927d9788dfa552797fb4555f21 not found: ID does not exist" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.012888 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:31:39 crc kubenswrapper[4744]: E1008 09:31:39.015460 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerName="nova-metadata-metadata" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.015489 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerName="nova-metadata-metadata" Oct 08 09:31:39 crc kubenswrapper[4744]: E1008 09:31:39.015511 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerName="nova-metadata-log" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.015518 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerName="nova-metadata-log" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.020181 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerName="nova-metadata-metadata" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.020223 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" containerName="nova-metadata-log" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.028148 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.028257 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.032899 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.033061 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.072540 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.072600 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-config-data\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.072665 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ssvs\" (UniqueName: \"kubernetes.io/projected/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-kube-api-access-5ssvs\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.072686 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.072708 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-logs\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.174874 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.174960 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-config-data\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.175023 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ssvs\" (UniqueName: \"kubernetes.io/projected/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-kube-api-access-5ssvs\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.175049 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.175661 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-logs\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.176106 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-logs\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.178631 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.179143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-config-data\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.180072 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.193925 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ssvs\" (UniqueName: \"kubernetes.io/projected/0592f7bc-38a5-43ad-ae49-acdb4f9023f4-kube-api-access-5ssvs\") pod \"nova-metadata-0\" (UID: \"0592f7bc-38a5-43ad-ae49-acdb4f9023f4\") " pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.352776 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.478337 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="237f9dff-1f92-46fa-a8d3-23f447f334be" path="/var/lib/kubelet/pods/237f9dff-1f92-46fa-a8d3-23f447f334be/volumes" Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.803052 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 09:31:39 crc kubenswrapper[4744]: W1008 09:31:39.821004 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0592f7bc_38a5_43ad_ae49_acdb4f9023f4.slice/crio-4d8204f9dc337adde7bef3a6bffaabe2bfe41f1d331692aa969fda1014db78ce WatchSource:0}: Error finding container 4d8204f9dc337adde7bef3a6bffaabe2bfe41f1d331692aa969fda1014db78ce: Status 404 returned error can't find the container with id 4d8204f9dc337adde7bef3a6bffaabe2bfe41f1d331692aa969fda1014db78ce Oct 08 09:31:39 crc kubenswrapper[4744]: I1008 09:31:39.952328 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0592f7bc-38a5-43ad-ae49-acdb4f9023f4","Type":"ContainerStarted","Data":"4d8204f9dc337adde7bef3a6bffaabe2bfe41f1d331692aa969fda1014db78ce"} Oct 08 09:31:40 crc kubenswrapper[4744]: I1008 09:31:40.971882 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0592f7bc-38a5-43ad-ae49-acdb4f9023f4","Type":"ContainerStarted","Data":"569bfcbfee0f1b1df35953d8544a60a6e02b10502e4e140ea51d197953ec2c1c"} Oct 08 09:31:40 crc kubenswrapper[4744]: I1008 09:31:40.972909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0592f7bc-38a5-43ad-ae49-acdb4f9023f4","Type":"ContainerStarted","Data":"824cb90484ef8a43dc4c699bd54f7dc8eab4c5f1876909ac3a085dc8ff906229"} Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.294707 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.777317 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.803582 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.803555822 podStartE2EDuration="4.803555822s" podCreationTimestamp="2025-10-08 09:31:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:31:41.004216176 +0000 UTC m=+1196.251861425" watchObservedRunningTime="2025-10-08 09:31:42.803555822 +0000 UTC m=+1198.051201061" Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.973337 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdb30ed4-01f7-4555-ac98-0ad496f7af40-logs\") pod \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.973769 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvr9l\" (UniqueName: \"kubernetes.io/projected/bdb30ed4-01f7-4555-ac98-0ad496f7af40-kube-api-access-pvr9l\") pod \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.973824 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-combined-ca-bundle\") pod \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.973852 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-internal-tls-certs\") pod \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.973941 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdb30ed4-01f7-4555-ac98-0ad496f7af40-logs" (OuterVolumeSpecName: "logs") pod "bdb30ed4-01f7-4555-ac98-0ad496f7af40" (UID: "bdb30ed4-01f7-4555-ac98-0ad496f7af40"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.974084 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-public-tls-certs\") pod \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.974124 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-config-data\") pod \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\" (UID: \"bdb30ed4-01f7-4555-ac98-0ad496f7af40\") " Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.975018 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdb30ed4-01f7-4555-ac98-0ad496f7af40-logs\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.989844 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdb30ed4-01f7-4555-ac98-0ad496f7af40-kube-api-access-pvr9l" (OuterVolumeSpecName: "kube-api-access-pvr9l") pod "bdb30ed4-01f7-4555-ac98-0ad496f7af40" (UID: "bdb30ed4-01f7-4555-ac98-0ad496f7af40"). InnerVolumeSpecName "kube-api-access-pvr9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.998636 4744 generic.go:334] "Generic (PLEG): container finished" podID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" containerID="47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be" exitCode=0 Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.998720 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bdb30ed4-01f7-4555-ac98-0ad496f7af40","Type":"ContainerDied","Data":"47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be"} Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.998752 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"bdb30ed4-01f7-4555-ac98-0ad496f7af40","Type":"ContainerDied","Data":"701827a267963e69cedd229954cc7c0a1b94cf69e1e7112c7166bfb1256c147a"} Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.998772 4744 scope.go:117] "RemoveContainer" containerID="47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be" Oct 08 09:31:42 crc kubenswrapper[4744]: I1008 09:31:42.998946 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.004938 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-config-data" (OuterVolumeSpecName: "config-data") pod "bdb30ed4-01f7-4555-ac98-0ad496f7af40" (UID: "bdb30ed4-01f7-4555-ac98-0ad496f7af40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.011466 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdb30ed4-01f7-4555-ac98-0ad496f7af40" (UID: "bdb30ed4-01f7-4555-ac98-0ad496f7af40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.055161 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bdb30ed4-01f7-4555-ac98-0ad496f7af40" (UID: "bdb30ed4-01f7-4555-ac98-0ad496f7af40"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.057112 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bdb30ed4-01f7-4555-ac98-0ad496f7af40" (UID: "bdb30ed4-01f7-4555-ac98-0ad496f7af40"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.076968 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.076999 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.077012 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvr9l\" (UniqueName: \"kubernetes.io/projected/bdb30ed4-01f7-4555-ac98-0ad496f7af40-kube-api-access-pvr9l\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.077022 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.077031 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdb30ed4-01f7-4555-ac98-0ad496f7af40-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.085666 4744 scope.go:117] "RemoveContainer" containerID="7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.101592 4744 scope.go:117] "RemoveContainer" containerID="47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be" Oct 08 09:31:43 crc kubenswrapper[4744]: E1008 09:31:43.101906 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be\": container with ID starting with 47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be not found: ID does not exist" containerID="47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.101936 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be"} err="failed to get container status \"47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be\": rpc error: code = NotFound desc = could not find container \"47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be\": container with ID starting with 47330e57bccf384838188c5a96b3c24164d72ee1b4c43e2a41c71a939df510be not found: ID does not exist" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.101963 4744 scope.go:117] "RemoveContainer" containerID="7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0" Oct 08 09:31:43 crc kubenswrapper[4744]: E1008 09:31:43.102172 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0\": container with ID starting with 7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0 not found: ID does not exist" containerID="7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.102191 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0"} err="failed to get container status \"7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0\": rpc error: code = NotFound desc = could not find container \"7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0\": container with ID starting with 7757e903109683386fbad17f1276b425145e14945caa79334a762529aafa9fe0 not found: ID does not exist" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.343446 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.350429 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.374919 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 09:31:43 crc kubenswrapper[4744]: E1008 09:31:43.375397 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" containerName="nova-api-api" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.375417 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" containerName="nova-api-api" Oct 08 09:31:43 crc kubenswrapper[4744]: E1008 09:31:43.375441 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" containerName="nova-api-log" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.375449 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" containerName="nova-api-log" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.375667 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" containerName="nova-api-api" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.375686 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" containerName="nova-api-log" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.376740 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.380863 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.381358 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.381502 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.390228 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63d92921-2a92-4233-8f69-02aa2c2b5984-internal-tls-certs\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.390309 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63d92921-2a92-4233-8f69-02aa2c2b5984-public-tls-certs\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.390344 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d92921-2a92-4233-8f69-02aa2c2b5984-logs\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.390367 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scvbd\" (UniqueName: \"kubernetes.io/projected/63d92921-2a92-4233-8f69-02aa2c2b5984-kube-api-access-scvbd\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.390424 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d92921-2a92-4233-8f69-02aa2c2b5984-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.390537 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d92921-2a92-4233-8f69-02aa2c2b5984-config-data\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.394731 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.471764 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdb30ed4-01f7-4555-ac98-0ad496f7af40" path="/var/lib/kubelet/pods/bdb30ed4-01f7-4555-ac98-0ad496f7af40/volumes" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.492413 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63d92921-2a92-4233-8f69-02aa2c2b5984-internal-tls-certs\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.492731 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63d92921-2a92-4233-8f69-02aa2c2b5984-public-tls-certs\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.492878 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d92921-2a92-4233-8f69-02aa2c2b5984-logs\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.492973 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scvbd\" (UniqueName: \"kubernetes.io/projected/63d92921-2a92-4233-8f69-02aa2c2b5984-kube-api-access-scvbd\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.493095 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d92921-2a92-4233-8f69-02aa2c2b5984-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.493544 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d92921-2a92-4233-8f69-02aa2c2b5984-config-data\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.494829 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63d92921-2a92-4233-8f69-02aa2c2b5984-logs\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.498988 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d92921-2a92-4233-8f69-02aa2c2b5984-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.501066 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63d92921-2a92-4233-8f69-02aa2c2b5984-config-data\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.510990 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63d92921-2a92-4233-8f69-02aa2c2b5984-internal-tls-certs\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.510996 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/63d92921-2a92-4233-8f69-02aa2c2b5984-public-tls-certs\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.518627 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scvbd\" (UniqueName: \"kubernetes.io/projected/63d92921-2a92-4233-8f69-02aa2c2b5984-kube-api-access-scvbd\") pod \"nova-api-0\" (UID: \"63d92921-2a92-4233-8f69-02aa2c2b5984\") " pod="openstack/nova-api-0" Oct 08 09:31:43 crc kubenswrapper[4744]: I1008 09:31:43.764771 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 09:31:44 crc kubenswrapper[4744]: W1008 09:31:44.244326 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63d92921_2a92_4233_8f69_02aa2c2b5984.slice/crio-cca558cecc0711b2bd684facf9876f33a3cc95fcbf96e4fdebdedacdc771b29b WatchSource:0}: Error finding container cca558cecc0711b2bd684facf9876f33a3cc95fcbf96e4fdebdedacdc771b29b: Status 404 returned error can't find the container with id cca558cecc0711b2bd684facf9876f33a3cc95fcbf96e4fdebdedacdc771b29b Oct 08 09:31:44 crc kubenswrapper[4744]: I1008 09:31:44.246763 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 09:31:44 crc kubenswrapper[4744]: I1008 09:31:44.353167 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 09:31:44 crc kubenswrapper[4744]: I1008 09:31:44.353254 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 09:31:45 crc kubenswrapper[4744]: I1008 09:31:45.040729 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"63d92921-2a92-4233-8f69-02aa2c2b5984","Type":"ContainerStarted","Data":"238e651ed8f4fbf60e89cd359e5bd3c3bb50aaea68e1681d614104fbda56d066"} Oct 08 09:31:45 crc kubenswrapper[4744]: I1008 09:31:45.041029 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"63d92921-2a92-4233-8f69-02aa2c2b5984","Type":"ContainerStarted","Data":"3c949d48522df48d05921a7a31fcec3eb29612e5b253c830a7298bfce2cd6c15"} Oct 08 09:31:45 crc kubenswrapper[4744]: I1008 09:31:45.041046 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"63d92921-2a92-4233-8f69-02aa2c2b5984","Type":"ContainerStarted","Data":"cca558cecc0711b2bd684facf9876f33a3cc95fcbf96e4fdebdedacdc771b29b"} Oct 08 09:31:45 crc kubenswrapper[4744]: I1008 09:31:45.065425 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.065402055 podStartE2EDuration="2.065402055s" podCreationTimestamp="2025-10-08 09:31:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:31:45.060299914 +0000 UTC m=+1200.307945153" watchObservedRunningTime="2025-10-08 09:31:45.065402055 +0000 UTC m=+1200.313047294" Oct 08 09:31:47 crc kubenswrapper[4744]: I1008 09:31:47.293390 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 09:31:47 crc kubenswrapper[4744]: I1008 09:31:47.330737 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 09:31:48 crc kubenswrapper[4744]: I1008 09:31:48.120185 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 09:31:49 crc kubenswrapper[4744]: I1008 09:31:49.353272 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 09:31:49 crc kubenswrapper[4744]: I1008 09:31:49.353653 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 09:31:50 crc kubenswrapper[4744]: I1008 09:31:50.365622 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0592f7bc-38a5-43ad-ae49-acdb4f9023f4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 09:31:50 crc kubenswrapper[4744]: I1008 09:31:50.365676 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0592f7bc-38a5-43ad-ae49-acdb4f9023f4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 09:31:53 crc kubenswrapper[4744]: I1008 09:31:53.024901 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 09:31:53 crc kubenswrapper[4744]: I1008 09:31:53.765407 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 09:31:53 crc kubenswrapper[4744]: I1008 09:31:53.765458 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 09:31:54 crc kubenswrapper[4744]: I1008 09:31:54.779538 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="63d92921-2a92-4233-8f69-02aa2c2b5984" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 09:31:54 crc kubenswrapper[4744]: I1008 09:31:54.780060 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="63d92921-2a92-4233-8f69-02aa2c2b5984" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.192:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 09:31:59 crc kubenswrapper[4744]: I1008 09:31:59.360186 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 09:31:59 crc kubenswrapper[4744]: I1008 09:31:59.362813 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 09:31:59 crc kubenswrapper[4744]: I1008 09:31:59.373405 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 09:32:00 crc kubenswrapper[4744]: I1008 09:32:00.299526 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 09:32:03 crc kubenswrapper[4744]: I1008 09:32:03.774730 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 09:32:03 crc kubenswrapper[4744]: I1008 09:32:03.778027 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 09:32:03 crc kubenswrapper[4744]: I1008 09:32:03.779620 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 09:32:03 crc kubenswrapper[4744]: I1008 09:32:03.802091 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 09:32:04 crc kubenswrapper[4744]: I1008 09:32:04.285577 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 09:32:04 crc kubenswrapper[4744]: I1008 09:32:04.313464 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 09:32:12 crc kubenswrapper[4744]: I1008 09:32:12.225911 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 09:32:13 crc kubenswrapper[4744]: I1008 09:32:13.582483 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 09:32:17 crc kubenswrapper[4744]: I1008 09:32:17.277555 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="6a6f5a1b-6abb-4568-89b5-a255f919cebb" containerName="rabbitmq" containerID="cri-o://ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72" gracePeriod=604795 Oct 08 09:32:18 crc kubenswrapper[4744]: I1008 09:32:18.862182 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" containerName="rabbitmq" containerID="cri-o://f3cb2f38c332b38d6da190637c53552d6176707980c4a0f214e672ede5b5273e" gracePeriod=604795 Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.831499 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.957547 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.957920 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-confd\") pod \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.957958 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-server-conf\") pod \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.958004 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-plugins-conf\") pod \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.958045 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-plugins\") pod \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.958274 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6a6f5a1b-6abb-4568-89b5-a255f919cebb-pod-info\") pod \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.958316 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-tls\") pod \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.958403 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbscx\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-kube-api-access-lbscx\") pod \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.958446 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-config-data\") pod \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.958487 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6a6f5a1b-6abb-4568-89b5-a255f919cebb-erlang-cookie-secret\") pod \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.958553 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-erlang-cookie\") pod \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\" (UID: \"6a6f5a1b-6abb-4568-89b5-a255f919cebb\") " Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.969599 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "6a6f5a1b-6abb-4568-89b5-a255f919cebb" (UID: "6a6f5a1b-6abb-4568-89b5-a255f919cebb"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.970539 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "6a6f5a1b-6abb-4568-89b5-a255f919cebb" (UID: "6a6f5a1b-6abb-4568-89b5-a255f919cebb"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.974901 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "6a6f5a1b-6abb-4568-89b5-a255f919cebb" (UID: "6a6f5a1b-6abb-4568-89b5-a255f919cebb"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.982694 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/6a6f5a1b-6abb-4568-89b5-a255f919cebb-pod-info" (OuterVolumeSpecName: "pod-info") pod "6a6f5a1b-6abb-4568-89b5-a255f919cebb" (UID: "6a6f5a1b-6abb-4568-89b5-a255f919cebb"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.987474 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "6a6f5a1b-6abb-4568-89b5-a255f919cebb" (UID: "6a6f5a1b-6abb-4568-89b5-a255f919cebb"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:32:23 crc kubenswrapper[4744]: I1008 09:32:23.996861 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-kube-api-access-lbscx" (OuterVolumeSpecName: "kube-api-access-lbscx") pod "6a6f5a1b-6abb-4568-89b5-a255f919cebb" (UID: "6a6f5a1b-6abb-4568-89b5-a255f919cebb"). InnerVolumeSpecName "kube-api-access-lbscx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.008840 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a6f5a1b-6abb-4568-89b5-a255f919cebb-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "6a6f5a1b-6abb-4568-89b5-a255f919cebb" (UID: "6a6f5a1b-6abb-4568-89b5-a255f919cebb"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.010784 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "6a6f5a1b-6abb-4568-89b5-a255f919cebb" (UID: "6a6f5a1b-6abb-4568-89b5-a255f919cebb"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.060855 4744 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6a6f5a1b-6abb-4568-89b5-a255f919cebb-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.060892 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.060919 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.060930 4744 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.060939 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.060946 4744 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6a6f5a1b-6abb-4568-89b5-a255f919cebb-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.060954 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.060961 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbscx\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-kube-api-access-lbscx\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.131434 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.160991 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-server-conf" (OuterVolumeSpecName: "server-conf") pod "6a6f5a1b-6abb-4568-89b5-a255f919cebb" (UID: "6a6f5a1b-6abb-4568-89b5-a255f919cebb"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.161498 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-config-data" (OuterVolumeSpecName: "config-data") pod "6a6f5a1b-6abb-4568-89b5-a255f919cebb" (UID: "6a6f5a1b-6abb-4568-89b5-a255f919cebb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.167093 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.167155 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.167187 4744 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6a6f5a1b-6abb-4568-89b5-a255f919cebb-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.169351 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "6a6f5a1b-6abb-4568-89b5-a255f919cebb" (UID: "6a6f5a1b-6abb-4568-89b5-a255f919cebb"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.269046 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6a6f5a1b-6abb-4568-89b5-a255f919cebb-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.477251 4744 generic.go:334] "Generic (PLEG): container finished" podID="6a6f5a1b-6abb-4568-89b5-a255f919cebb" containerID="ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72" exitCode=0 Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.477308 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6a6f5a1b-6abb-4568-89b5-a255f919cebb","Type":"ContainerDied","Data":"ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72"} Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.477352 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6a6f5a1b-6abb-4568-89b5-a255f919cebb","Type":"ContainerDied","Data":"eeedf4ac70685ed31ef68628b235f1aca550e7054155460c642b17603761d187"} Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.477393 4744 scope.go:117] "RemoveContainer" containerID="ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.478559 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.498837 4744 scope.go:117] "RemoveContainer" containerID="816a11709d28ae06d2f534850a7bff8d19e598d0723d36416765dd4b97d1863e" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.552872 4744 scope.go:117] "RemoveContainer" containerID="ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72" Oct 08 09:32:24 crc kubenswrapper[4744]: E1008 09:32:24.553541 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72\": container with ID starting with ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72 not found: ID does not exist" containerID="ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.553629 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72"} err="failed to get container status \"ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72\": rpc error: code = NotFound desc = could not find container \"ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72\": container with ID starting with ea124b968e48fa240e14c6a77cc24feb28060ded68166e74ee56103343e2ed72 not found: ID does not exist" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.553677 4744 scope.go:117] "RemoveContainer" containerID="816a11709d28ae06d2f534850a7bff8d19e598d0723d36416765dd4b97d1863e" Oct 08 09:32:24 crc kubenswrapper[4744]: E1008 09:32:24.554321 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"816a11709d28ae06d2f534850a7bff8d19e598d0723d36416765dd4b97d1863e\": container with ID starting with 816a11709d28ae06d2f534850a7bff8d19e598d0723d36416765dd4b97d1863e not found: ID does not exist" containerID="816a11709d28ae06d2f534850a7bff8d19e598d0723d36416765dd4b97d1863e" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.554355 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"816a11709d28ae06d2f534850a7bff8d19e598d0723d36416765dd4b97d1863e"} err="failed to get container status \"816a11709d28ae06d2f534850a7bff8d19e598d0723d36416765dd4b97d1863e\": rpc error: code = NotFound desc = could not find container \"816a11709d28ae06d2f534850a7bff8d19e598d0723d36416765dd4b97d1863e\": container with ID starting with 816a11709d28ae06d2f534850a7bff8d19e598d0723d36416765dd4b97d1863e not found: ID does not exist" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.604334 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.615311 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.634189 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 09:32:24 crc kubenswrapper[4744]: E1008 09:32:24.634760 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a6f5a1b-6abb-4568-89b5-a255f919cebb" containerName="setup-container" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.634776 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a6f5a1b-6abb-4568-89b5-a255f919cebb" containerName="setup-container" Oct 08 09:32:24 crc kubenswrapper[4744]: E1008 09:32:24.634805 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a6f5a1b-6abb-4568-89b5-a255f919cebb" containerName="rabbitmq" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.634813 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a6f5a1b-6abb-4568-89b5-a255f919cebb" containerName="rabbitmq" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.635004 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a6f5a1b-6abb-4568-89b5-a255f919cebb" containerName="rabbitmq" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.636184 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.643825 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.646404 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.646612 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.646811 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.647107 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5qvmw" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.647271 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.648292 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.673649 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.780492 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/19186b52-dfab-4d1f-9ccd-176abc9975ff-pod-info\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.780896 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/19186b52-dfab-4d1f-9ccd-176abc9975ff-server-conf\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.780935 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/19186b52-dfab-4d1f-9ccd-176abc9975ff-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.781001 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvc7q\" (UniqueName: \"kubernetes.io/projected/19186b52-dfab-4d1f-9ccd-176abc9975ff-kube-api-access-rvc7q\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.781022 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/19186b52-dfab-4d1f-9ccd-176abc9975ff-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.781040 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/19186b52-dfab-4d1f-9ccd-176abc9975ff-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.781059 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/19186b52-dfab-4d1f-9ccd-176abc9975ff-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.781090 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/19186b52-dfab-4d1f-9ccd-176abc9975ff-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.781109 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/19186b52-dfab-4d1f-9ccd-176abc9975ff-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.781129 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19186b52-dfab-4d1f-9ccd-176abc9975ff-config-data\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.781158 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.882843 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/19186b52-dfab-4d1f-9ccd-176abc9975ff-server-conf\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.882902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/19186b52-dfab-4d1f-9ccd-176abc9975ff-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.882967 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvc7q\" (UniqueName: \"kubernetes.io/projected/19186b52-dfab-4d1f-9ccd-176abc9975ff-kube-api-access-rvc7q\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.882988 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/19186b52-dfab-4d1f-9ccd-176abc9975ff-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.883008 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/19186b52-dfab-4d1f-9ccd-176abc9975ff-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.883026 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/19186b52-dfab-4d1f-9ccd-176abc9975ff-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.883055 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/19186b52-dfab-4d1f-9ccd-176abc9975ff-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.883076 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/19186b52-dfab-4d1f-9ccd-176abc9975ff-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.883098 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19186b52-dfab-4d1f-9ccd-176abc9975ff-config-data\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.883124 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.883150 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/19186b52-dfab-4d1f-9ccd-176abc9975ff-pod-info\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.883532 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/19186b52-dfab-4d1f-9ccd-176abc9975ff-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.883643 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/19186b52-dfab-4d1f-9ccd-176abc9975ff-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.883951 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.884239 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19186b52-dfab-4d1f-9ccd-176abc9975ff-config-data\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.884939 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/19186b52-dfab-4d1f-9ccd-176abc9975ff-server-conf\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.886193 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/19186b52-dfab-4d1f-9ccd-176abc9975ff-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.888509 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/19186b52-dfab-4d1f-9ccd-176abc9975ff-pod-info\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.888856 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/19186b52-dfab-4d1f-9ccd-176abc9975ff-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.898135 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/19186b52-dfab-4d1f-9ccd-176abc9975ff-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.900144 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/19186b52-dfab-4d1f-9ccd-176abc9975ff-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.913203 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvc7q\" (UniqueName: \"kubernetes.io/projected/19186b52-dfab-4d1f-9ccd-176abc9975ff-kube-api-access-rvc7q\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.950316 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"19186b52-dfab-4d1f-9ccd-176abc9975ff\") " pod="openstack/rabbitmq-server-0" Oct 08 09:32:24 crc kubenswrapper[4744]: I1008 09:32:24.962723 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.470157 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a6f5a1b-6abb-4568-89b5-a255f919cebb" path="/var/lib/kubelet/pods/6a6f5a1b-6abb-4568-89b5-a255f919cebb/volumes" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.491454 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 09:32:25 crc kubenswrapper[4744]: W1008 09:32:25.495292 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19186b52_dfab_4d1f_9ccd_176abc9975ff.slice/crio-455fa084f2f1bb1310867c7f4aae7bb871bc2209f5d84692bc216fa19ab51b55 WatchSource:0}: Error finding container 455fa084f2f1bb1310867c7f4aae7bb871bc2209f5d84692bc216fa19ab51b55: Status 404 returned error can't find the container with id 455fa084f2f1bb1310867c7f4aae7bb871bc2209f5d84692bc216fa19ab51b55 Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.495515 4744 generic.go:334] "Generic (PLEG): container finished" podID="18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" containerID="f3cb2f38c332b38d6da190637c53552d6176707980c4a0f214e672ede5b5273e" exitCode=0 Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.495596 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6","Type":"ContainerDied","Data":"f3cb2f38c332b38d6da190637c53552d6176707980c4a0f214e672ede5b5273e"} Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.719748 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.812785 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dql57\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-kube-api-access-dql57\") pod \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.812878 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-confd\") pod \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.812980 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-erlang-cookie-secret\") pod \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.813005 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-tls\") pod \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.813151 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-plugins\") pod \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.813322 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.813397 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-config-data\") pod \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.813418 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-pod-info\") pod \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.813476 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-plugins-conf\") pod \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.813511 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-server-conf\") pod \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.813925 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-erlang-cookie\") pod \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\" (UID: \"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6\") " Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.829727 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" (UID: "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.831860 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" (UID: "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.831941 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" (UID: "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.832146 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" (UID: "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.835060 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-kube-api-access-dql57" (OuterVolumeSpecName: "kube-api-access-dql57") pod "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" (UID: "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6"). InnerVolumeSpecName "kube-api-access-dql57". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.837119 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" (UID: "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.838140 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" (UID: "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.860639 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-pod-info" (OuterVolumeSpecName: "pod-info") pod "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" (UID: "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.891467 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-config-data" (OuterVolumeSpecName: "config-data") pod "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" (UID: "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.914518 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-server-conf" (OuterVolumeSpecName: "server-conf") pod "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" (UID: "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.923856 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dql57\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-kube-api-access-dql57\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.923905 4744 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.923923 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.923935 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.923964 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.923976 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.923987 4744 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.923998 4744 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.924008 4744 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.924021 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:25 crc kubenswrapper[4744]: I1008 09:32:25.961317 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.002570 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" (UID: "18a53e2c-fcfa-4f49-aef0-ef1cee4118f6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.027969 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.028022 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.532707 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.532705 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18a53e2c-fcfa-4f49-aef0-ef1cee4118f6","Type":"ContainerDied","Data":"1df0e399ccf38a5de5ca1c8c40ff27077c3c6be4db4c989dbfa297e8dc1d8dab"} Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.533250 4744 scope.go:117] "RemoveContainer" containerID="f3cb2f38c332b38d6da190637c53552d6176707980c4a0f214e672ede5b5273e" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.534995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"19186b52-dfab-4d1f-9ccd-176abc9975ff","Type":"ContainerStarted","Data":"455fa084f2f1bb1310867c7f4aae7bb871bc2209f5d84692bc216fa19ab51b55"} Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.642160 4744 scope.go:117] "RemoveContainer" containerID="a7ae4524f517636f791ec2332601150d328e8d3eb0890dfd3c61265b8ed54e20" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.652769 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.671809 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.689213 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 09:32:26 crc kubenswrapper[4744]: E1008 09:32:26.693797 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" containerName="rabbitmq" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.693824 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" containerName="rabbitmq" Oct 08 09:32:26 crc kubenswrapper[4744]: E1008 09:32:26.693852 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" containerName="setup-container" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.693864 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" containerName="setup-container" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.694058 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" containerName="rabbitmq" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.697087 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.703641 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.703641 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5sp9s" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.703952 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.704033 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.705665 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.706354 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.707259 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.729165 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.850217 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc9mj\" (UniqueName: \"kubernetes.io/projected/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-kube-api-access-cc9mj\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.850338 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.850404 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.850443 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.850519 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.850562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.850599 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.850653 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.850695 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.850775 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.850796 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.952983 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.953116 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.953913 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.955335 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.954020 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.955767 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.956112 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.956399 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.956508 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.956539 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.956589 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.956774 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.956790 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.956842 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc9mj\" (UniqueName: \"kubernetes.io/projected/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-kube-api-access-cc9mj\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.957760 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.957810 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.958505 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.960573 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.961107 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.961183 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.961713 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.984776 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc9mj\" (UniqueName: \"kubernetes.io/projected/8ae09ae5-011d-48f9-ad83-a53bfb91c90c-kube-api-access-cc9mj\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:26 crc kubenswrapper[4744]: I1008 09:32:26.988909 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8ae09ae5-011d-48f9-ad83-a53bfb91c90c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:27 crc kubenswrapper[4744]: I1008 09:32:27.019948 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:32:27 crc kubenswrapper[4744]: I1008 09:32:27.468297 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18a53e2c-fcfa-4f49-aef0-ef1cee4118f6" path="/var/lib/kubelet/pods/18a53e2c-fcfa-4f49-aef0-ef1cee4118f6/volumes" Oct 08 09:32:27 crc kubenswrapper[4744]: I1008 09:32:27.561132 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"19186b52-dfab-4d1f-9ccd-176abc9975ff","Type":"ContainerStarted","Data":"23dba2a9b822644072a616ad09ad1fcb6c31a514896f04042b5e6893a192ed6b"} Oct 08 09:32:27 crc kubenswrapper[4744]: I1008 09:32:27.563678 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8ae09ae5-011d-48f9-ad83-a53bfb91c90c","Type":"ContainerStarted","Data":"04cbf7dd76a8c068277ca4e6a2da49f3e4d6a3eeca1a23046ec91ebd9ccb644d"} Oct 08 09:32:27 crc kubenswrapper[4744]: I1008 09:32:27.573902 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.590891 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8ae09ae5-011d-48f9-ad83-a53bfb91c90c","Type":"ContainerStarted","Data":"cfc78ca625438ea1acfbff2330637920cdf607b6b20f0502a42d5eed2a542e68"} Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.766171 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64fb5d8fd7-lqhc4"] Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.767758 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.789949 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.816592 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64fb5d8fd7-lqhc4"] Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.834950 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-config\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.835087 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdjpf\" (UniqueName: \"kubernetes.io/projected/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-kube-api-access-gdjpf\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.835169 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-ovsdbserver-sb\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.835254 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-dns-svc\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.835350 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-ovsdbserver-nb\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.835418 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-openstack-edpm-ipam\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.942337 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdjpf\" (UniqueName: \"kubernetes.io/projected/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-kube-api-access-gdjpf\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.942452 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-ovsdbserver-sb\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.942516 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-dns-svc\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.942595 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-ovsdbserver-nb\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.942638 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-openstack-edpm-ipam\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.942755 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-config\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.943958 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-config\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.949834 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-ovsdbserver-sb\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.950400 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-dns-svc\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.950929 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-ovsdbserver-nb\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.951435 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-openstack-edpm-ipam\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:29 crc kubenswrapper[4744]: I1008 09:32:29.987002 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdjpf\" (UniqueName: \"kubernetes.io/projected/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-kube-api-access-gdjpf\") pod \"dnsmasq-dns-64fb5d8fd7-lqhc4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:30 crc kubenswrapper[4744]: I1008 09:32:30.121780 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:31 crc kubenswrapper[4744]: I1008 09:32:31.367929 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64fb5d8fd7-lqhc4"] Oct 08 09:32:31 crc kubenswrapper[4744]: I1008 09:32:31.614246 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" event={"ID":"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4","Type":"ContainerStarted","Data":"5c230db9bf1bde9a9926b37f0f1c2001a64a4bf83fe52c52b104cb45ae136da4"} Oct 08 09:32:31 crc kubenswrapper[4744]: I1008 09:32:31.614323 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" event={"ID":"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4","Type":"ContainerStarted","Data":"fffd2d2fb78af65b26bc5d8751a0687b19a98ad570ee17d4c3b85858baf49b7b"} Oct 08 09:32:32 crc kubenswrapper[4744]: I1008 09:32:32.625407 4744 generic.go:334] "Generic (PLEG): container finished" podID="9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" containerID="5c230db9bf1bde9a9926b37f0f1c2001a64a4bf83fe52c52b104cb45ae136da4" exitCode=0 Oct 08 09:32:32 crc kubenswrapper[4744]: I1008 09:32:32.625610 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" event={"ID":"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4","Type":"ContainerDied","Data":"5c230db9bf1bde9a9926b37f0f1c2001a64a4bf83fe52c52b104cb45ae136da4"} Oct 08 09:32:33 crc kubenswrapper[4744]: I1008 09:32:33.636728 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" event={"ID":"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4","Type":"ContainerStarted","Data":"e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9"} Oct 08 09:32:33 crc kubenswrapper[4744]: I1008 09:32:33.637933 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:33 crc kubenswrapper[4744]: I1008 09:32:33.667821 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" podStartSLOduration=4.667801708 podStartE2EDuration="4.667801708s" podCreationTimestamp="2025-10-08 09:32:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:32:33.662955775 +0000 UTC m=+1248.910601014" watchObservedRunningTime="2025-10-08 09:32:33.667801708 +0000 UTC m=+1248.915446947" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.124245 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.220428 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-665946c669-qmkg4"] Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.220659 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-665946c669-qmkg4" podUID="3d6356c2-49f8-46b8-a5cc-6aae66240e2e" containerName="dnsmasq-dns" containerID="cri-o://b3a7a841f7e53db91817f6d3673002b5aa6c516f49d951f043223801430cbeed" gracePeriod=10 Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.459225 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-867c8fd5c5-wkswc"] Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.461865 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.510408 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-867c8fd5c5-wkswc"] Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.588131 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw822\" (UniqueName: \"kubernetes.io/projected/65ab44b6-3478-42d3-b9cb-64f911f13342-kube-api-access-dw822\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.588204 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-ovsdbserver-sb\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.588277 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-ovsdbserver-nb\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.588356 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-dns-svc\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.588398 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-config\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.588413 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-openstack-edpm-ipam\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.702867 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-ovsdbserver-nb\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.703389 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-dns-svc\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.703423 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-config\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.703441 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-openstack-edpm-ipam\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.703530 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw822\" (UniqueName: \"kubernetes.io/projected/65ab44b6-3478-42d3-b9cb-64f911f13342-kube-api-access-dw822\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.703585 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-ovsdbserver-sb\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.704555 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-dns-svc\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.704610 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-ovsdbserver-sb\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.705257 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-openstack-edpm-ipam\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.707076 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-ovsdbserver-nb\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.708935 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-config\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.718774 4744 generic.go:334] "Generic (PLEG): container finished" podID="3d6356c2-49f8-46b8-a5cc-6aae66240e2e" containerID="b3a7a841f7e53db91817f6d3673002b5aa6c516f49d951f043223801430cbeed" exitCode=0 Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.718834 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665946c669-qmkg4" event={"ID":"3d6356c2-49f8-46b8-a5cc-6aae66240e2e","Type":"ContainerDied","Data":"b3a7a841f7e53db91817f6d3673002b5aa6c516f49d951f043223801430cbeed"} Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.732032 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw822\" (UniqueName: \"kubernetes.io/projected/65ab44b6-3478-42d3-b9cb-64f911f13342-kube-api-access-dw822\") pod \"dnsmasq-dns-867c8fd5c5-wkswc\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.788286 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:40 crc kubenswrapper[4744]: I1008 09:32:40.922984 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.009821 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-ovsdbserver-nb\") pod \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.010144 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-ovsdbserver-sb\") pod \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.010190 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpsgf\" (UniqueName: \"kubernetes.io/projected/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-kube-api-access-dpsgf\") pod \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.010260 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-config\") pod \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.010285 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-dns-svc\") pod \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\" (UID: \"3d6356c2-49f8-46b8-a5cc-6aae66240e2e\") " Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.017809 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-kube-api-access-dpsgf" (OuterVolumeSpecName: "kube-api-access-dpsgf") pod "3d6356c2-49f8-46b8-a5cc-6aae66240e2e" (UID: "3d6356c2-49f8-46b8-a5cc-6aae66240e2e"). InnerVolumeSpecName "kube-api-access-dpsgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.078917 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-config" (OuterVolumeSpecName: "config") pod "3d6356c2-49f8-46b8-a5cc-6aae66240e2e" (UID: "3d6356c2-49f8-46b8-a5cc-6aae66240e2e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.084411 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3d6356c2-49f8-46b8-a5cc-6aae66240e2e" (UID: "3d6356c2-49f8-46b8-a5cc-6aae66240e2e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.099433 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3d6356c2-49f8-46b8-a5cc-6aae66240e2e" (UID: "3d6356c2-49f8-46b8-a5cc-6aae66240e2e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.112698 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.112729 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpsgf\" (UniqueName: \"kubernetes.io/projected/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-kube-api-access-dpsgf\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.112741 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.112750 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.117754 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3d6356c2-49f8-46b8-a5cc-6aae66240e2e" (UID: "3d6356c2-49f8-46b8-a5cc-6aae66240e2e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.215628 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d6356c2-49f8-46b8-a5cc-6aae66240e2e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.340472 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-867c8fd5c5-wkswc"] Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.731648 4744 generic.go:334] "Generic (PLEG): container finished" podID="65ab44b6-3478-42d3-b9cb-64f911f13342" containerID="38f1147a1a639caf9a23723086d64ebcda76d5cd7c811418cdea6642b1368e6e" exitCode=0 Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.731725 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" event={"ID":"65ab44b6-3478-42d3-b9cb-64f911f13342","Type":"ContainerDied","Data":"38f1147a1a639caf9a23723086d64ebcda76d5cd7c811418cdea6642b1368e6e"} Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.732006 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" event={"ID":"65ab44b6-3478-42d3-b9cb-64f911f13342","Type":"ContainerStarted","Data":"43b6ad8279ddf9acaed3feeeb223bde6e75cd46c627f21f2e95b393501fcabdf"} Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.735104 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-665946c669-qmkg4" event={"ID":"3d6356c2-49f8-46b8-a5cc-6aae66240e2e","Type":"ContainerDied","Data":"d82723004149c36ae1131c8bead029150cb577b19c90bfa662f94a86871d798e"} Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.735165 4744 scope.go:117] "RemoveContainer" containerID="b3a7a841f7e53db91817f6d3673002b5aa6c516f49d951f043223801430cbeed" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.735398 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-665946c669-qmkg4" Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.791722 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-665946c669-qmkg4"] Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.798953 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-665946c669-qmkg4"] Oct 08 09:32:41 crc kubenswrapper[4744]: I1008 09:32:41.861638 4744 scope.go:117] "RemoveContainer" containerID="758b3491697cca71f5668ef8a725447d4f6cd2f9a79f86722fbd89463c05259c" Oct 08 09:32:42 crc kubenswrapper[4744]: I1008 09:32:42.748911 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" event={"ID":"65ab44b6-3478-42d3-b9cb-64f911f13342","Type":"ContainerStarted","Data":"754ac2af91ca329df2d3d3340340cb5a2294499e668c5b00a0374f7e45b216dc"} Oct 08 09:32:42 crc kubenswrapper[4744]: I1008 09:32:42.749521 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:42 crc kubenswrapper[4744]: I1008 09:32:42.775391 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" podStartSLOduration=2.7753501050000002 podStartE2EDuration="2.775350105s" podCreationTimestamp="2025-10-08 09:32:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:32:42.769700519 +0000 UTC m=+1258.017345768" watchObservedRunningTime="2025-10-08 09:32:42.775350105 +0000 UTC m=+1258.022995344" Oct 08 09:32:43 crc kubenswrapper[4744]: I1008 09:32:43.466348 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d6356c2-49f8-46b8-a5cc-6aae66240e2e" path="/var/lib/kubelet/pods/3d6356c2-49f8-46b8-a5cc-6aae66240e2e/volumes" Oct 08 09:32:50 crc kubenswrapper[4744]: I1008 09:32:50.790713 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 09:32:50 crc kubenswrapper[4744]: I1008 09:32:50.908007 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64fb5d8fd7-lqhc4"] Oct 08 09:32:50 crc kubenswrapper[4744]: I1008 09:32:50.908460 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" podUID="9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" containerName="dnsmasq-dns" containerID="cri-o://e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9" gracePeriod=10 Oct 08 09:32:51 crc kubenswrapper[4744]: I1008 09:32:51.937144 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:51 crc kubenswrapper[4744]: I1008 09:32:51.971804 4744 generic.go:334] "Generic (PLEG): container finished" podID="9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" containerID="e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9" exitCode=0 Oct 08 09:32:51 crc kubenswrapper[4744]: I1008 09:32:51.971884 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" event={"ID":"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4","Type":"ContainerDied","Data":"e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9"} Oct 08 09:32:51 crc kubenswrapper[4744]: I1008 09:32:51.971925 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" event={"ID":"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4","Type":"ContainerDied","Data":"fffd2d2fb78af65b26bc5d8751a0687b19a98ad570ee17d4c3b85858baf49b7b"} Oct 08 09:32:51 crc kubenswrapper[4744]: I1008 09:32:51.971947 4744 scope.go:117] "RemoveContainer" containerID="e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9" Oct 08 09:32:51 crc kubenswrapper[4744]: I1008 09:32:51.972124 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64fb5d8fd7-lqhc4" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:51.999966 4744 scope.go:117] "RemoveContainer" containerID="5c230db9bf1bde9a9926b37f0f1c2001a64a4bf83fe52c52b104cb45ae136da4" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.034702 4744 scope.go:117] "RemoveContainer" containerID="e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9" Oct 08 09:32:52 crc kubenswrapper[4744]: E1008 09:32:52.036044 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9\": container with ID starting with e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9 not found: ID does not exist" containerID="e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.036116 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9"} err="failed to get container status \"e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9\": rpc error: code = NotFound desc = could not find container \"e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9\": container with ID starting with e955fbdaf70bc36e2868c33b5f45f38ae92aac09e9f7461d1a7d779a8c7b39b9 not found: ID does not exist" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.036164 4744 scope.go:117] "RemoveContainer" containerID="5c230db9bf1bde9a9926b37f0f1c2001a64a4bf83fe52c52b104cb45ae136da4" Oct 08 09:32:52 crc kubenswrapper[4744]: E1008 09:32:52.036988 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c230db9bf1bde9a9926b37f0f1c2001a64a4bf83fe52c52b104cb45ae136da4\": container with ID starting with 5c230db9bf1bde9a9926b37f0f1c2001a64a4bf83fe52c52b104cb45ae136da4 not found: ID does not exist" containerID="5c230db9bf1bde9a9926b37f0f1c2001a64a4bf83fe52c52b104cb45ae136da4" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.037021 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c230db9bf1bde9a9926b37f0f1c2001a64a4bf83fe52c52b104cb45ae136da4"} err="failed to get container status \"5c230db9bf1bde9a9926b37f0f1c2001a64a4bf83fe52c52b104cb45ae136da4\": rpc error: code = NotFound desc = could not find container \"5c230db9bf1bde9a9926b37f0f1c2001a64a4bf83fe52c52b104cb45ae136da4\": container with ID starting with 5c230db9bf1bde9a9926b37f0f1c2001a64a4bf83fe52c52b104cb45ae136da4 not found: ID does not exist" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.109461 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-openstack-edpm-ipam\") pod \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.109513 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-ovsdbserver-nb\") pod \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.109554 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-config\") pod \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.109583 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-dns-svc\") pod \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.109776 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdjpf\" (UniqueName: \"kubernetes.io/projected/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-kube-api-access-gdjpf\") pod \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.109886 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-ovsdbserver-sb\") pod \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\" (UID: \"9592f2b6-43b2-4cd4-b808-52bf4e03e3e4\") " Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.118667 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-kube-api-access-gdjpf" (OuterVolumeSpecName: "kube-api-access-gdjpf") pod "9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" (UID: "9592f2b6-43b2-4cd4-b808-52bf4e03e3e4"). InnerVolumeSpecName "kube-api-access-gdjpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.182000 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" (UID: "9592f2b6-43b2-4cd4-b808-52bf4e03e3e4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.187313 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-config" (OuterVolumeSpecName: "config") pod "9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" (UID: "9592f2b6-43b2-4cd4-b808-52bf4e03e3e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.194178 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" (UID: "9592f2b6-43b2-4cd4-b808-52bf4e03e3e4"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.212015 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" (UID: "9592f2b6-43b2-4cd4-b808-52bf4e03e3e4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.212548 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdjpf\" (UniqueName: \"kubernetes.io/projected/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-kube-api-access-gdjpf\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.212590 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.212603 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.212616 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-config\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.212626 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.221508 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" (UID: "9592f2b6-43b2-4cd4-b808-52bf4e03e3e4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.315890 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.326422 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64fb5d8fd7-lqhc4"] Oct 08 09:32:52 crc kubenswrapper[4744]: I1008 09:32:52.336278 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64fb5d8fd7-lqhc4"] Oct 08 09:32:53 crc kubenswrapper[4744]: I1008 09:32:53.467859 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" path="/var/lib/kubelet/pods/9592f2b6-43b2-4cd4-b808-52bf4e03e3e4/volumes" Oct 08 09:33:00 crc kubenswrapper[4744]: I1008 09:33:00.074469 4744 generic.go:334] "Generic (PLEG): container finished" podID="19186b52-dfab-4d1f-9ccd-176abc9975ff" containerID="23dba2a9b822644072a616ad09ad1fcb6c31a514896f04042b5e6893a192ed6b" exitCode=0 Oct 08 09:33:00 crc kubenswrapper[4744]: I1008 09:33:00.074521 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"19186b52-dfab-4d1f-9ccd-176abc9975ff","Type":"ContainerDied","Data":"23dba2a9b822644072a616ad09ad1fcb6c31a514896f04042b5e6893a192ed6b"} Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.096431 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"19186b52-dfab-4d1f-9ccd-176abc9975ff","Type":"ContainerStarted","Data":"1dafbcb441efb19e094fe1708555b8d8bab08746707e6367190efbaa1f56ce97"} Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.097200 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.105354 4744 generic.go:334] "Generic (PLEG): container finished" podID="8ae09ae5-011d-48f9-ad83-a53bfb91c90c" containerID="cfc78ca625438ea1acfbff2330637920cdf607b6b20f0502a42d5eed2a542e68" exitCode=0 Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.107833 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8ae09ae5-011d-48f9-ad83-a53bfb91c90c","Type":"ContainerDied","Data":"cfc78ca625438ea1acfbff2330637920cdf607b6b20f0502a42d5eed2a542e68"} Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.136005 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv"] Oct 08 09:33:01 crc kubenswrapper[4744]: E1008 09:33:01.137004 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" containerName="init" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.137111 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" containerName="init" Oct 08 09:33:01 crc kubenswrapper[4744]: E1008 09:33:01.137170 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" containerName="dnsmasq-dns" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.137223 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" containerName="dnsmasq-dns" Oct 08 09:33:01 crc kubenswrapper[4744]: E1008 09:33:01.137290 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d6356c2-49f8-46b8-a5cc-6aae66240e2e" containerName="dnsmasq-dns" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.137348 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d6356c2-49f8-46b8-a5cc-6aae66240e2e" containerName="dnsmasq-dns" Oct 08 09:33:01 crc kubenswrapper[4744]: E1008 09:33:01.137492 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d6356c2-49f8-46b8-a5cc-6aae66240e2e" containerName="init" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.137555 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d6356c2-49f8-46b8-a5cc-6aae66240e2e" containerName="init" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.137835 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9592f2b6-43b2-4cd4-b808-52bf4e03e3e4" containerName="dnsmasq-dns" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.138008 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d6356c2-49f8-46b8-a5cc-6aae66240e2e" containerName="dnsmasq-dns" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.138978 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.154218 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.154737 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.154822 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.154994 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.161652 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv"] Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.171485 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.171466887 podStartE2EDuration="37.171466887s" podCreationTimestamp="2025-10-08 09:32:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:33:01.168747482 +0000 UTC m=+1276.416392741" watchObservedRunningTime="2025-10-08 09:33:01.171466887 +0000 UTC m=+1276.419112126" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.211936 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbmg6\" (UniqueName: \"kubernetes.io/projected/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-kube-api-access-zbmg6\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.212247 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.214840 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.220198 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: E1008 09:33:01.258600 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ae09ae5_011d_48f9_ad83_a53bfb91c90c.slice/crio-conmon-cfc78ca625438ea1acfbff2330637920cdf607b6b20f0502a42d5eed2a542e68.scope\": RecentStats: unable to find data in memory cache]" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.324480 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.324584 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbmg6\" (UniqueName: \"kubernetes.io/projected/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-kube-api-access-zbmg6\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.324624 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.324660 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.330865 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.331427 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.338210 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.367734 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbmg6\" (UniqueName: \"kubernetes.io/projected/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-kube-api-access-zbmg6\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:01 crc kubenswrapper[4744]: I1008 09:33:01.659204 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:02 crc kubenswrapper[4744]: I1008 09:33:02.121121 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8ae09ae5-011d-48f9-ad83-a53bfb91c90c","Type":"ContainerStarted","Data":"e2e82ef071f2af5100cb61fa79a16cc9a53ce92ed00052a5843496cc33d994b9"} Oct 08 09:33:02 crc kubenswrapper[4744]: I1008 09:33:02.122228 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:33:02 crc kubenswrapper[4744]: I1008 09:33:02.162322 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.16230149 podStartE2EDuration="36.16230149s" podCreationTimestamp="2025-10-08 09:32:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 09:33:02.156928882 +0000 UTC m=+1277.404574131" watchObservedRunningTime="2025-10-08 09:33:02.16230149 +0000 UTC m=+1277.409946729" Oct 08 09:33:02 crc kubenswrapper[4744]: I1008 09:33:02.329453 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv"] Oct 08 09:33:02 crc kubenswrapper[4744]: W1008 09:33:02.345844 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4b8727b_2bfd_4e9d_839d_8c7e69925c9b.slice/crio-ec83c107013db0c285bb201c17c43767d60a5d1bcd52b326e91ed6b044d74c8c WatchSource:0}: Error finding container ec83c107013db0c285bb201c17c43767d60a5d1bcd52b326e91ed6b044d74c8c: Status 404 returned error can't find the container with id ec83c107013db0c285bb201c17c43767d60a5d1bcd52b326e91ed6b044d74c8c Oct 08 09:33:03 crc kubenswrapper[4744]: I1008 09:33:03.130163 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" event={"ID":"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b","Type":"ContainerStarted","Data":"ec83c107013db0c285bb201c17c43767d60a5d1bcd52b326e91ed6b044d74c8c"} Oct 08 09:33:14 crc kubenswrapper[4744]: I1008 09:33:14.966402 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="19186b52-dfab-4d1f-9ccd-176abc9975ff" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.193:5671: connect: connection refused" Oct 08 09:33:17 crc kubenswrapper[4744]: I1008 09:33:17.022688 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="8ae09ae5-011d-48f9-ad83-a53bfb91c90c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.194:5671: connect: connection refused" Oct 08 09:33:19 crc kubenswrapper[4744]: I1008 09:33:19.336951 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" event={"ID":"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b","Type":"ContainerStarted","Data":"8aa46e31f25cdc651f531c404f2421a2c12c7db27284eb48602697e74262cf42"} Oct 08 09:33:19 crc kubenswrapper[4744]: I1008 09:33:19.375624 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" podStartSLOduration=2.324274683 podStartE2EDuration="18.375599764s" podCreationTimestamp="2025-10-08 09:33:01 +0000 UTC" firstStartedPulling="2025-10-08 09:33:02.347682844 +0000 UTC m=+1277.595328093" lastFinishedPulling="2025-10-08 09:33:18.399007935 +0000 UTC m=+1293.646653174" observedRunningTime="2025-10-08 09:33:19.369394223 +0000 UTC m=+1294.617039462" watchObservedRunningTime="2025-10-08 09:33:19.375599764 +0000 UTC m=+1294.623245003" Oct 08 09:33:19 crc kubenswrapper[4744]: I1008 09:33:19.690077 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:33:19 crc kubenswrapper[4744]: I1008 09:33:19.690142 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:33:24 crc kubenswrapper[4744]: I1008 09:33:24.965098 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 08 09:33:27 crc kubenswrapper[4744]: I1008 09:33:27.023687 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 08 09:33:32 crc kubenswrapper[4744]: I1008 09:33:32.453965 4744 generic.go:334] "Generic (PLEG): container finished" podID="e4b8727b-2bfd-4e9d-839d-8c7e69925c9b" containerID="8aa46e31f25cdc651f531c404f2421a2c12c7db27284eb48602697e74262cf42" exitCode=0 Oct 08 09:33:32 crc kubenswrapper[4744]: I1008 09:33:32.455009 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" event={"ID":"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b","Type":"ContainerDied","Data":"8aa46e31f25cdc651f531c404f2421a2c12c7db27284eb48602697e74262cf42"} Oct 08 09:33:33 crc kubenswrapper[4744]: I1008 09:33:33.892567 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.057184 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbmg6\" (UniqueName: \"kubernetes.io/projected/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-kube-api-access-zbmg6\") pod \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.057342 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-ssh-key\") pod \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.057745 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-repo-setup-combined-ca-bundle\") pod \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.057812 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-inventory\") pod \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\" (UID: \"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b\") " Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.064298 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "e4b8727b-2bfd-4e9d-839d-8c7e69925c9b" (UID: "e4b8727b-2bfd-4e9d-839d-8c7e69925c9b"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.065252 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-kube-api-access-zbmg6" (OuterVolumeSpecName: "kube-api-access-zbmg6") pod "e4b8727b-2bfd-4e9d-839d-8c7e69925c9b" (UID: "e4b8727b-2bfd-4e9d-839d-8c7e69925c9b"). InnerVolumeSpecName "kube-api-access-zbmg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.099900 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-inventory" (OuterVolumeSpecName: "inventory") pod "e4b8727b-2bfd-4e9d-839d-8c7e69925c9b" (UID: "e4b8727b-2bfd-4e9d-839d-8c7e69925c9b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.108790 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e4b8727b-2bfd-4e9d-839d-8c7e69925c9b" (UID: "e4b8727b-2bfd-4e9d-839d-8c7e69925c9b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.161652 4744 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.161684 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.161694 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbmg6\" (UniqueName: \"kubernetes.io/projected/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-kube-api-access-zbmg6\") on node \"crc\" DevicePath \"\"" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.161703 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.479049 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" event={"ID":"e4b8727b-2bfd-4e9d-839d-8c7e69925c9b","Type":"ContainerDied","Data":"ec83c107013db0c285bb201c17c43767d60a5d1bcd52b326e91ed6b044d74c8c"} Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.479101 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec83c107013db0c285bb201c17c43767d60a5d1bcd52b326e91ed6b044d74c8c" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.479169 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.569333 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz"] Oct 08 09:33:34 crc kubenswrapper[4744]: E1008 09:33:34.569856 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4b8727b-2bfd-4e9d-839d-8c7e69925c9b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.569876 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4b8727b-2bfd-4e9d-839d-8c7e69925c9b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.570057 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4b8727b-2bfd-4e9d-839d-8c7e69925c9b" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.570879 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.573614 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.573720 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.574308 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.577401 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.587812 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz"] Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.672866 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.673007 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.673044 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvwkb\" (UniqueName: \"kubernetes.io/projected/9511d654-cff2-45ce-9b6a-a3d98744e0a6-kube-api-access-dvwkb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.673104 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.775060 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.775155 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.775196 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvwkb\" (UniqueName: \"kubernetes.io/projected/9511d654-cff2-45ce-9b6a-a3d98744e0a6-kube-api-access-dvwkb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.775275 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.780785 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.783778 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.783835 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.796623 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvwkb\" (UniqueName: \"kubernetes.io/projected/9511d654-cff2-45ce-9b6a-a3d98744e0a6-kube-api-access-dvwkb\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:34 crc kubenswrapper[4744]: I1008 09:33:34.932194 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:33:35 crc kubenswrapper[4744]: I1008 09:33:35.512924 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz"] Oct 08 09:33:36 crc kubenswrapper[4744]: I1008 09:33:36.501572 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" event={"ID":"9511d654-cff2-45ce-9b6a-a3d98744e0a6","Type":"ContainerStarted","Data":"1b67b47da4a52585a88d6873c6a6f7a8bb324cb24c8579f5bb9a0fc23cbaf42e"} Oct 08 09:33:36 crc kubenswrapper[4744]: I1008 09:33:36.502096 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" event={"ID":"9511d654-cff2-45ce-9b6a-a3d98744e0a6","Type":"ContainerStarted","Data":"3e0ee4acc03a6f0692e31795a6ffc471a9d125e33dccc60648b9cd2f8cc64b4b"} Oct 08 09:33:36 crc kubenswrapper[4744]: I1008 09:33:36.534077 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" podStartSLOduration=2.066428746 podStartE2EDuration="2.534050487s" podCreationTimestamp="2025-10-08 09:33:34 +0000 UTC" firstStartedPulling="2025-10-08 09:33:35.527973946 +0000 UTC m=+1310.775619185" lastFinishedPulling="2025-10-08 09:33:35.995595687 +0000 UTC m=+1311.243240926" observedRunningTime="2025-10-08 09:33:36.526619906 +0000 UTC m=+1311.774265145" watchObservedRunningTime="2025-10-08 09:33:36.534050487 +0000 UTC m=+1311.781695726" Oct 08 09:33:49 crc kubenswrapper[4744]: I1008 09:33:49.690695 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:33:49 crc kubenswrapper[4744]: I1008 09:33:49.691728 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:33:52 crc kubenswrapper[4744]: I1008 09:33:52.775492 4744 scope.go:117] "RemoveContainer" containerID="2fd80a6c6a013ebf4855e67c1c9e44f84a6e75f18ffcae45276348043008e93c" Oct 08 09:33:52 crc kubenswrapper[4744]: I1008 09:33:52.826641 4744 scope.go:117] "RemoveContainer" containerID="02160fdf25f738b58ca129e0f0a8927722ca3ba72b1ea0fc0371d083ba9304d5" Oct 08 09:34:19 crc kubenswrapper[4744]: I1008 09:34:19.690662 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:34:19 crc kubenswrapper[4744]: I1008 09:34:19.691171 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:34:19 crc kubenswrapper[4744]: I1008 09:34:19.691436 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:34:19 crc kubenswrapper[4744]: I1008 09:34:19.692621 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"936d206815e6d287935cf3c047d48aada7c4fce45edc4c83541f33b9de7f8702"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 09:34:19 crc kubenswrapper[4744]: I1008 09:34:19.692689 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://936d206815e6d287935cf3c047d48aada7c4fce45edc4c83541f33b9de7f8702" gracePeriod=600 Oct 08 09:34:19 crc kubenswrapper[4744]: I1008 09:34:19.974289 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="936d206815e6d287935cf3c047d48aada7c4fce45edc4c83541f33b9de7f8702" exitCode=0 Oct 08 09:34:19 crc kubenswrapper[4744]: I1008 09:34:19.974350 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"936d206815e6d287935cf3c047d48aada7c4fce45edc4c83541f33b9de7f8702"} Oct 08 09:34:19 crc kubenswrapper[4744]: I1008 09:34:19.974512 4744 scope.go:117] "RemoveContainer" containerID="fc794dea8745685aa845397cb02ef8c3c9ae63ea691d5f60cd28e1c653036510" Oct 08 09:34:20 crc kubenswrapper[4744]: I1008 09:34:20.992478 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb"} Oct 08 09:34:52 crc kubenswrapper[4744]: I1008 09:34:52.929988 4744 scope.go:117] "RemoveContainer" containerID="4a732de79ee5e3345c73aa7ae53e7b04d65649de5902781c1a2937bf75137412" Oct 08 09:34:52 crc kubenswrapper[4744]: I1008 09:34:52.954755 4744 scope.go:117] "RemoveContainer" containerID="0e72b9aa009c9fbe7c47cf736ce7c073904482761dba03f427c8bb92ab81380e" Oct 08 09:34:53 crc kubenswrapper[4744]: I1008 09:34:53.010102 4744 scope.go:117] "RemoveContainer" containerID="798489ddd7442f1d55d1aea17083ea9445108c87809654cc94d54563aa808f0c" Oct 08 09:35:59 crc kubenswrapper[4744]: I1008 09:35:59.830212 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z2lgt"] Oct 08 09:35:59 crc kubenswrapper[4744]: I1008 09:35:59.838690 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:35:59 crc kubenswrapper[4744]: I1008 09:35:59.859170 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z2lgt"] Oct 08 09:35:59 crc kubenswrapper[4744]: I1008 09:35:59.988077 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56068c3-9b81-4c8e-9934-67858fff01cf-utilities\") pod \"community-operators-z2lgt\" (UID: \"d56068c3-9b81-4c8e-9934-67858fff01cf\") " pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:35:59 crc kubenswrapper[4744]: I1008 09:35:59.988651 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxhbq\" (UniqueName: \"kubernetes.io/projected/d56068c3-9b81-4c8e-9934-67858fff01cf-kube-api-access-fxhbq\") pod \"community-operators-z2lgt\" (UID: \"d56068c3-9b81-4c8e-9934-67858fff01cf\") " pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:35:59 crc kubenswrapper[4744]: I1008 09:35:59.988992 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56068c3-9b81-4c8e-9934-67858fff01cf-catalog-content\") pod \"community-operators-z2lgt\" (UID: \"d56068c3-9b81-4c8e-9934-67858fff01cf\") " pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:00 crc kubenswrapper[4744]: I1008 09:36:00.090897 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56068c3-9b81-4c8e-9934-67858fff01cf-catalog-content\") pod \"community-operators-z2lgt\" (UID: \"d56068c3-9b81-4c8e-9934-67858fff01cf\") " pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:00 crc kubenswrapper[4744]: I1008 09:36:00.090991 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56068c3-9b81-4c8e-9934-67858fff01cf-utilities\") pod \"community-operators-z2lgt\" (UID: \"d56068c3-9b81-4c8e-9934-67858fff01cf\") " pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:00 crc kubenswrapper[4744]: I1008 09:36:00.091107 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxhbq\" (UniqueName: \"kubernetes.io/projected/d56068c3-9b81-4c8e-9934-67858fff01cf-kube-api-access-fxhbq\") pod \"community-operators-z2lgt\" (UID: \"d56068c3-9b81-4c8e-9934-67858fff01cf\") " pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:00 crc kubenswrapper[4744]: I1008 09:36:00.091400 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56068c3-9b81-4c8e-9934-67858fff01cf-catalog-content\") pod \"community-operators-z2lgt\" (UID: \"d56068c3-9b81-4c8e-9934-67858fff01cf\") " pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:00 crc kubenswrapper[4744]: I1008 09:36:00.091486 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56068c3-9b81-4c8e-9934-67858fff01cf-utilities\") pod \"community-operators-z2lgt\" (UID: \"d56068c3-9b81-4c8e-9934-67858fff01cf\") " pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:00 crc kubenswrapper[4744]: I1008 09:36:00.117579 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxhbq\" (UniqueName: \"kubernetes.io/projected/d56068c3-9b81-4c8e-9934-67858fff01cf-kube-api-access-fxhbq\") pod \"community-operators-z2lgt\" (UID: \"d56068c3-9b81-4c8e-9934-67858fff01cf\") " pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:00 crc kubenswrapper[4744]: I1008 09:36:00.162880 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:00 crc kubenswrapper[4744]: I1008 09:36:00.642614 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z2lgt"] Oct 08 09:36:00 crc kubenswrapper[4744]: I1008 09:36:00.966435 4744 generic.go:334] "Generic (PLEG): container finished" podID="d56068c3-9b81-4c8e-9934-67858fff01cf" containerID="35452fee091f45f2781e743c9eb9324efc34fd09bfdc26cc5bb09ca056dd1e60" exitCode=0 Oct 08 09:36:00 crc kubenswrapper[4744]: I1008 09:36:00.966489 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2lgt" event={"ID":"d56068c3-9b81-4c8e-9934-67858fff01cf","Type":"ContainerDied","Data":"35452fee091f45f2781e743c9eb9324efc34fd09bfdc26cc5bb09ca056dd1e60"} Oct 08 09:36:00 crc kubenswrapper[4744]: I1008 09:36:00.966518 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2lgt" event={"ID":"d56068c3-9b81-4c8e-9934-67858fff01cf","Type":"ContainerStarted","Data":"81607da5bf4545162bf5a42376516a4ede8b0f2cfbdd08063f483ce9a54b1bb7"} Oct 08 09:36:01 crc kubenswrapper[4744]: I1008 09:36:01.988291 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2lgt" event={"ID":"d56068c3-9b81-4c8e-9934-67858fff01cf","Type":"ContainerStarted","Data":"1f3221ad741b7dbe80cae0d0a4bcc3b7e008fb232ef0fb0f0e3feedd6ff17927"} Oct 08 09:36:04 crc kubenswrapper[4744]: I1008 09:36:04.020067 4744 generic.go:334] "Generic (PLEG): container finished" podID="d56068c3-9b81-4c8e-9934-67858fff01cf" containerID="1f3221ad741b7dbe80cae0d0a4bcc3b7e008fb232ef0fb0f0e3feedd6ff17927" exitCode=0 Oct 08 09:36:04 crc kubenswrapper[4744]: I1008 09:36:04.020163 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2lgt" event={"ID":"d56068c3-9b81-4c8e-9934-67858fff01cf","Type":"ContainerDied","Data":"1f3221ad741b7dbe80cae0d0a4bcc3b7e008fb232ef0fb0f0e3feedd6ff17927"} Oct 08 09:36:05 crc kubenswrapper[4744]: I1008 09:36:05.055871 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2lgt" event={"ID":"d56068c3-9b81-4c8e-9934-67858fff01cf","Type":"ContainerStarted","Data":"df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962"} Oct 08 09:36:05 crc kubenswrapper[4744]: I1008 09:36:05.084787 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z2lgt" podStartSLOduration=2.583964957 podStartE2EDuration="6.08476445s" podCreationTimestamp="2025-10-08 09:35:59 +0000 UTC" firstStartedPulling="2025-10-08 09:36:00.968013665 +0000 UTC m=+1456.215658904" lastFinishedPulling="2025-10-08 09:36:04.468813148 +0000 UTC m=+1459.716458397" observedRunningTime="2025-10-08 09:36:05.075121752 +0000 UTC m=+1460.322767011" watchObservedRunningTime="2025-10-08 09:36:05.08476445 +0000 UTC m=+1460.332409689" Oct 08 09:36:10 crc kubenswrapper[4744]: I1008 09:36:10.163440 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:10 crc kubenswrapper[4744]: I1008 09:36:10.164015 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:10 crc kubenswrapper[4744]: I1008 09:36:10.215657 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:11 crc kubenswrapper[4744]: I1008 09:36:11.164867 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:11 crc kubenswrapper[4744]: I1008 09:36:11.216516 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z2lgt"] Oct 08 09:36:13 crc kubenswrapper[4744]: I1008 09:36:13.133567 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z2lgt" podUID="d56068c3-9b81-4c8e-9934-67858fff01cf" containerName="registry-server" containerID="cri-o://df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962" gracePeriod=2 Oct 08 09:36:13 crc kubenswrapper[4744]: I1008 09:36:13.780590 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:13 crc kubenswrapper[4744]: I1008 09:36:13.960320 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxhbq\" (UniqueName: \"kubernetes.io/projected/d56068c3-9b81-4c8e-9934-67858fff01cf-kube-api-access-fxhbq\") pod \"d56068c3-9b81-4c8e-9934-67858fff01cf\" (UID: \"d56068c3-9b81-4c8e-9934-67858fff01cf\") " Oct 08 09:36:13 crc kubenswrapper[4744]: I1008 09:36:13.961271 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56068c3-9b81-4c8e-9934-67858fff01cf-utilities\") pod \"d56068c3-9b81-4c8e-9934-67858fff01cf\" (UID: \"d56068c3-9b81-4c8e-9934-67858fff01cf\") " Oct 08 09:36:13 crc kubenswrapper[4744]: I1008 09:36:13.961392 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56068c3-9b81-4c8e-9934-67858fff01cf-catalog-content\") pod \"d56068c3-9b81-4c8e-9934-67858fff01cf\" (UID: \"d56068c3-9b81-4c8e-9934-67858fff01cf\") " Oct 08 09:36:13 crc kubenswrapper[4744]: I1008 09:36:13.962530 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d56068c3-9b81-4c8e-9934-67858fff01cf-utilities" (OuterVolumeSpecName: "utilities") pod "d56068c3-9b81-4c8e-9934-67858fff01cf" (UID: "d56068c3-9b81-4c8e-9934-67858fff01cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:36:13 crc kubenswrapper[4744]: I1008 09:36:13.970733 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d56068c3-9b81-4c8e-9934-67858fff01cf-kube-api-access-fxhbq" (OuterVolumeSpecName: "kube-api-access-fxhbq") pod "d56068c3-9b81-4c8e-9934-67858fff01cf" (UID: "d56068c3-9b81-4c8e-9934-67858fff01cf"). InnerVolumeSpecName "kube-api-access-fxhbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.031184 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d56068c3-9b81-4c8e-9934-67858fff01cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d56068c3-9b81-4c8e-9934-67858fff01cf" (UID: "d56068c3-9b81-4c8e-9934-67858fff01cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.064564 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxhbq\" (UniqueName: \"kubernetes.io/projected/d56068c3-9b81-4c8e-9934-67858fff01cf-kube-api-access-fxhbq\") on node \"crc\" DevicePath \"\"" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.064928 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56068c3-9b81-4c8e-9934-67858fff01cf-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.064943 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56068c3-9b81-4c8e-9934-67858fff01cf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.145837 4744 generic.go:334] "Generic (PLEG): container finished" podID="d56068c3-9b81-4c8e-9934-67858fff01cf" containerID="df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962" exitCode=0 Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.145904 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2lgt" event={"ID":"d56068c3-9b81-4c8e-9934-67858fff01cf","Type":"ContainerDied","Data":"df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962"} Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.146569 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z2lgt" event={"ID":"d56068c3-9b81-4c8e-9934-67858fff01cf","Type":"ContainerDied","Data":"81607da5bf4545162bf5a42376516a4ede8b0f2cfbdd08063f483ce9a54b1bb7"} Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.146599 4744 scope.go:117] "RemoveContainer" containerID="df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.145928 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z2lgt" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.203734 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z2lgt"] Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.209636 4744 scope.go:117] "RemoveContainer" containerID="1f3221ad741b7dbe80cae0d0a4bcc3b7e008fb232ef0fb0f0e3feedd6ff17927" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.218592 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z2lgt"] Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.234508 4744 scope.go:117] "RemoveContainer" containerID="35452fee091f45f2781e743c9eb9324efc34fd09bfdc26cc5bb09ca056dd1e60" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.275840 4744 scope.go:117] "RemoveContainer" containerID="df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962" Oct 08 09:36:14 crc kubenswrapper[4744]: E1008 09:36:14.276464 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962\": container with ID starting with df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962 not found: ID does not exist" containerID="df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.276514 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962"} err="failed to get container status \"df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962\": rpc error: code = NotFound desc = could not find container \"df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962\": container with ID starting with df88702992875b91adab1bb8de74ee8972f23b33a2f0b2a55fca3bc28ff18962 not found: ID does not exist" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.276552 4744 scope.go:117] "RemoveContainer" containerID="1f3221ad741b7dbe80cae0d0a4bcc3b7e008fb232ef0fb0f0e3feedd6ff17927" Oct 08 09:36:14 crc kubenswrapper[4744]: E1008 09:36:14.277207 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f3221ad741b7dbe80cae0d0a4bcc3b7e008fb232ef0fb0f0e3feedd6ff17927\": container with ID starting with 1f3221ad741b7dbe80cae0d0a4bcc3b7e008fb232ef0fb0f0e3feedd6ff17927 not found: ID does not exist" containerID="1f3221ad741b7dbe80cae0d0a4bcc3b7e008fb232ef0fb0f0e3feedd6ff17927" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.277249 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f3221ad741b7dbe80cae0d0a4bcc3b7e008fb232ef0fb0f0e3feedd6ff17927"} err="failed to get container status \"1f3221ad741b7dbe80cae0d0a4bcc3b7e008fb232ef0fb0f0e3feedd6ff17927\": rpc error: code = NotFound desc = could not find container \"1f3221ad741b7dbe80cae0d0a4bcc3b7e008fb232ef0fb0f0e3feedd6ff17927\": container with ID starting with 1f3221ad741b7dbe80cae0d0a4bcc3b7e008fb232ef0fb0f0e3feedd6ff17927 not found: ID does not exist" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.277282 4744 scope.go:117] "RemoveContainer" containerID="35452fee091f45f2781e743c9eb9324efc34fd09bfdc26cc5bb09ca056dd1e60" Oct 08 09:36:14 crc kubenswrapper[4744]: E1008 09:36:14.277582 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35452fee091f45f2781e743c9eb9324efc34fd09bfdc26cc5bb09ca056dd1e60\": container with ID starting with 35452fee091f45f2781e743c9eb9324efc34fd09bfdc26cc5bb09ca056dd1e60 not found: ID does not exist" containerID="35452fee091f45f2781e743c9eb9324efc34fd09bfdc26cc5bb09ca056dd1e60" Oct 08 09:36:14 crc kubenswrapper[4744]: I1008 09:36:14.277614 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35452fee091f45f2781e743c9eb9324efc34fd09bfdc26cc5bb09ca056dd1e60"} err="failed to get container status \"35452fee091f45f2781e743c9eb9324efc34fd09bfdc26cc5bb09ca056dd1e60\": rpc error: code = NotFound desc = could not find container \"35452fee091f45f2781e743c9eb9324efc34fd09bfdc26cc5bb09ca056dd1e60\": container with ID starting with 35452fee091f45f2781e743c9eb9324efc34fd09bfdc26cc5bb09ca056dd1e60 not found: ID does not exist" Oct 08 09:36:15 crc kubenswrapper[4744]: I1008 09:36:15.462558 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d56068c3-9b81-4c8e-9934-67858fff01cf" path="/var/lib/kubelet/pods/d56068c3-9b81-4c8e-9934-67858fff01cf/volumes" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.129081 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n47gm"] Oct 08 09:36:18 crc kubenswrapper[4744]: E1008 09:36:18.130210 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56068c3-9b81-4c8e-9934-67858fff01cf" containerName="extract-utilities" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.130228 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56068c3-9b81-4c8e-9934-67858fff01cf" containerName="extract-utilities" Oct 08 09:36:18 crc kubenswrapper[4744]: E1008 09:36:18.130251 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56068c3-9b81-4c8e-9934-67858fff01cf" containerName="registry-server" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.130262 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56068c3-9b81-4c8e-9934-67858fff01cf" containerName="registry-server" Oct 08 09:36:18 crc kubenswrapper[4744]: E1008 09:36:18.130329 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56068c3-9b81-4c8e-9934-67858fff01cf" containerName="extract-content" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.130341 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56068c3-9b81-4c8e-9934-67858fff01cf" containerName="extract-content" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.130655 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d56068c3-9b81-4c8e-9934-67858fff01cf" containerName="registry-server" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.132661 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.148038 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n47gm"] Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.252417 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-utilities\") pod \"redhat-marketplace-n47gm\" (UID: \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\") " pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.252738 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-catalog-content\") pod \"redhat-marketplace-n47gm\" (UID: \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\") " pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.252974 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn8g2\" (UniqueName: \"kubernetes.io/projected/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-kube-api-access-cn8g2\") pod \"redhat-marketplace-n47gm\" (UID: \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\") " pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.354725 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-utilities\") pod \"redhat-marketplace-n47gm\" (UID: \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\") " pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.354864 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-catalog-content\") pod \"redhat-marketplace-n47gm\" (UID: \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\") " pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.354947 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn8g2\" (UniqueName: \"kubernetes.io/projected/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-kube-api-access-cn8g2\") pod \"redhat-marketplace-n47gm\" (UID: \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\") " pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.355301 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-utilities\") pod \"redhat-marketplace-n47gm\" (UID: \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\") " pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.355306 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-catalog-content\") pod \"redhat-marketplace-n47gm\" (UID: \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\") " pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.385643 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn8g2\" (UniqueName: \"kubernetes.io/projected/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-kube-api-access-cn8g2\") pod \"redhat-marketplace-n47gm\" (UID: \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\") " pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.459331 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:18 crc kubenswrapper[4744]: I1008 09:36:18.991870 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n47gm"] Oct 08 09:36:19 crc kubenswrapper[4744]: I1008 09:36:19.202761 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n47gm" event={"ID":"d2c1c075-2b01-4ce4-b851-9d15ee9afa57","Type":"ContainerStarted","Data":"babf5c31d2750a5462cb590235b48521b505ee244b1f915f0eecf9e4d59fcf3c"} Oct 08 09:36:20 crc kubenswrapper[4744]: I1008 09:36:20.214562 4744 generic.go:334] "Generic (PLEG): container finished" podID="d2c1c075-2b01-4ce4-b851-9d15ee9afa57" containerID="feee463a970858c5556e507243fbb37eb60b7c4f9f46f0494cb9ca7779c56572" exitCode=0 Oct 08 09:36:20 crc kubenswrapper[4744]: I1008 09:36:20.214676 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n47gm" event={"ID":"d2c1c075-2b01-4ce4-b851-9d15ee9afa57","Type":"ContainerDied","Data":"feee463a970858c5556e507243fbb37eb60b7c4f9f46f0494cb9ca7779c56572"} Oct 08 09:36:21 crc kubenswrapper[4744]: I1008 09:36:21.229412 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n47gm" event={"ID":"d2c1c075-2b01-4ce4-b851-9d15ee9afa57","Type":"ContainerStarted","Data":"50d851a16d63d4920ff87f49b9e0f0e422770a7ece92afbd3638a398e180fa78"} Oct 08 09:36:22 crc kubenswrapper[4744]: I1008 09:36:22.243594 4744 generic.go:334] "Generic (PLEG): container finished" podID="d2c1c075-2b01-4ce4-b851-9d15ee9afa57" containerID="50d851a16d63d4920ff87f49b9e0f0e422770a7ece92afbd3638a398e180fa78" exitCode=0 Oct 08 09:36:22 crc kubenswrapper[4744]: I1008 09:36:22.243659 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n47gm" event={"ID":"d2c1c075-2b01-4ce4-b851-9d15ee9afa57","Type":"ContainerDied","Data":"50d851a16d63d4920ff87f49b9e0f0e422770a7ece92afbd3638a398e180fa78"} Oct 08 09:36:23 crc kubenswrapper[4744]: I1008 09:36:23.258950 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n47gm" event={"ID":"d2c1c075-2b01-4ce4-b851-9d15ee9afa57","Type":"ContainerStarted","Data":"895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730"} Oct 08 09:36:23 crc kubenswrapper[4744]: I1008 09:36:23.304730 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n47gm" podStartSLOduration=2.865826817 podStartE2EDuration="5.304697898s" podCreationTimestamp="2025-10-08 09:36:18 +0000 UTC" firstStartedPulling="2025-10-08 09:36:20.217847556 +0000 UTC m=+1475.465492795" lastFinishedPulling="2025-10-08 09:36:22.656718637 +0000 UTC m=+1477.904363876" observedRunningTime="2025-10-08 09:36:23.297519995 +0000 UTC m=+1478.545165254" watchObservedRunningTime="2025-10-08 09:36:23.304697898 +0000 UTC m=+1478.552343137" Oct 08 09:36:28 crc kubenswrapper[4744]: I1008 09:36:28.459973 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:28 crc kubenswrapper[4744]: I1008 09:36:28.461444 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:28 crc kubenswrapper[4744]: I1008 09:36:28.512709 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:29 crc kubenswrapper[4744]: I1008 09:36:29.431677 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:29 crc kubenswrapper[4744]: I1008 09:36:29.511022 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n47gm"] Oct 08 09:36:31 crc kubenswrapper[4744]: I1008 09:36:31.388603 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n47gm" podUID="d2c1c075-2b01-4ce4-b851-9d15ee9afa57" containerName="registry-server" containerID="cri-o://895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730" gracePeriod=2 Oct 08 09:36:31 crc kubenswrapper[4744]: I1008 09:36:31.857382 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:31 crc kubenswrapper[4744]: I1008 09:36:31.941904 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-utilities\") pod \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\" (UID: \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\") " Oct 08 09:36:31 crc kubenswrapper[4744]: I1008 09:36:31.942067 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-catalog-content\") pod \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\" (UID: \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\") " Oct 08 09:36:31 crc kubenswrapper[4744]: I1008 09:36:31.942402 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn8g2\" (UniqueName: \"kubernetes.io/projected/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-kube-api-access-cn8g2\") pod \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\" (UID: \"d2c1c075-2b01-4ce4-b851-9d15ee9afa57\") " Oct 08 09:36:31 crc kubenswrapper[4744]: I1008 09:36:31.943144 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-utilities" (OuterVolumeSpecName: "utilities") pod "d2c1c075-2b01-4ce4-b851-9d15ee9afa57" (UID: "d2c1c075-2b01-4ce4-b851-9d15ee9afa57"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:36:31 crc kubenswrapper[4744]: I1008 09:36:31.947821 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-kube-api-access-cn8g2" (OuterVolumeSpecName: "kube-api-access-cn8g2") pod "d2c1c075-2b01-4ce4-b851-9d15ee9afa57" (UID: "d2c1c075-2b01-4ce4-b851-9d15ee9afa57"). InnerVolumeSpecName "kube-api-access-cn8g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:36:31 crc kubenswrapper[4744]: I1008 09:36:31.956410 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2c1c075-2b01-4ce4-b851-9d15ee9afa57" (UID: "d2c1c075-2b01-4ce4-b851-9d15ee9afa57"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.045114 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn8g2\" (UniqueName: \"kubernetes.io/projected/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-kube-api-access-cn8g2\") on node \"crc\" DevicePath \"\"" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.045181 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.045196 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2c1c075-2b01-4ce4-b851-9d15ee9afa57-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.403000 4744 generic.go:334] "Generic (PLEG): container finished" podID="d2c1c075-2b01-4ce4-b851-9d15ee9afa57" containerID="895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730" exitCode=0 Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.403052 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n47gm" event={"ID":"d2c1c075-2b01-4ce4-b851-9d15ee9afa57","Type":"ContainerDied","Data":"895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730"} Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.403080 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n47gm" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.403099 4744 scope.go:117] "RemoveContainer" containerID="895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.403085 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n47gm" event={"ID":"d2c1c075-2b01-4ce4-b851-9d15ee9afa57","Type":"ContainerDied","Data":"babf5c31d2750a5462cb590235b48521b505ee244b1f915f0eecf9e4d59fcf3c"} Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.451104 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n47gm"] Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.451792 4744 scope.go:117] "RemoveContainer" containerID="50d851a16d63d4920ff87f49b9e0f0e422770a7ece92afbd3638a398e180fa78" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.460676 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n47gm"] Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.481776 4744 scope.go:117] "RemoveContainer" containerID="feee463a970858c5556e507243fbb37eb60b7c4f9f46f0494cb9ca7779c56572" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.530627 4744 scope.go:117] "RemoveContainer" containerID="895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730" Oct 08 09:36:32 crc kubenswrapper[4744]: E1008 09:36:32.531109 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730\": container with ID starting with 895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730 not found: ID does not exist" containerID="895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.531145 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730"} err="failed to get container status \"895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730\": rpc error: code = NotFound desc = could not find container \"895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730\": container with ID starting with 895e7f794afb1faccced2ef0a4993ac3cc5916843c2400b242286ecb7aa95730 not found: ID does not exist" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.531170 4744 scope.go:117] "RemoveContainer" containerID="50d851a16d63d4920ff87f49b9e0f0e422770a7ece92afbd3638a398e180fa78" Oct 08 09:36:32 crc kubenswrapper[4744]: E1008 09:36:32.531598 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50d851a16d63d4920ff87f49b9e0f0e422770a7ece92afbd3638a398e180fa78\": container with ID starting with 50d851a16d63d4920ff87f49b9e0f0e422770a7ece92afbd3638a398e180fa78 not found: ID does not exist" containerID="50d851a16d63d4920ff87f49b9e0f0e422770a7ece92afbd3638a398e180fa78" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.531623 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50d851a16d63d4920ff87f49b9e0f0e422770a7ece92afbd3638a398e180fa78"} err="failed to get container status \"50d851a16d63d4920ff87f49b9e0f0e422770a7ece92afbd3638a398e180fa78\": rpc error: code = NotFound desc = could not find container \"50d851a16d63d4920ff87f49b9e0f0e422770a7ece92afbd3638a398e180fa78\": container with ID starting with 50d851a16d63d4920ff87f49b9e0f0e422770a7ece92afbd3638a398e180fa78 not found: ID does not exist" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.531635 4744 scope.go:117] "RemoveContainer" containerID="feee463a970858c5556e507243fbb37eb60b7c4f9f46f0494cb9ca7779c56572" Oct 08 09:36:32 crc kubenswrapper[4744]: E1008 09:36:32.531847 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feee463a970858c5556e507243fbb37eb60b7c4f9f46f0494cb9ca7779c56572\": container with ID starting with feee463a970858c5556e507243fbb37eb60b7c4f9f46f0494cb9ca7779c56572 not found: ID does not exist" containerID="feee463a970858c5556e507243fbb37eb60b7c4f9f46f0494cb9ca7779c56572" Oct 08 09:36:32 crc kubenswrapper[4744]: I1008 09:36:32.531869 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feee463a970858c5556e507243fbb37eb60b7c4f9f46f0494cb9ca7779c56572"} err="failed to get container status \"feee463a970858c5556e507243fbb37eb60b7c4f9f46f0494cb9ca7779c56572\": rpc error: code = NotFound desc = could not find container \"feee463a970858c5556e507243fbb37eb60b7c4f9f46f0494cb9ca7779c56572\": container with ID starting with feee463a970858c5556e507243fbb37eb60b7c4f9f46f0494cb9ca7779c56572 not found: ID does not exist" Oct 08 09:36:33 crc kubenswrapper[4744]: I1008 09:36:33.466575 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2c1c075-2b01-4ce4-b851-9d15ee9afa57" path="/var/lib/kubelet/pods/d2c1c075-2b01-4ce4-b851-9d15ee9afa57/volumes" Oct 08 09:36:49 crc kubenswrapper[4744]: I1008 09:36:49.690911 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:36:49 crc kubenswrapper[4744]: I1008 09:36:49.691696 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:36:53 crc kubenswrapper[4744]: I1008 09:36:53.135037 4744 scope.go:117] "RemoveContainer" containerID="652899e54dcc6b2ad13dd42267f9a8e10cb6f8f9f0b0d585511ea9005450af3c" Oct 08 09:36:53 crc kubenswrapper[4744]: I1008 09:36:53.163637 4744 scope.go:117] "RemoveContainer" containerID="cc71f0c8c72c15736d655cfbd3aae35933ef71569a56dff8053d12bb6374e1fa" Oct 08 09:36:58 crc kubenswrapper[4744]: I1008 09:36:58.675974 4744 generic.go:334] "Generic (PLEG): container finished" podID="9511d654-cff2-45ce-9b6a-a3d98744e0a6" containerID="1b67b47da4a52585a88d6873c6a6f7a8bb324cb24c8579f5bb9a0fc23cbaf42e" exitCode=0 Oct 08 09:36:58 crc kubenswrapper[4744]: I1008 09:36:58.676064 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" event={"ID":"9511d654-cff2-45ce-9b6a-a3d98744e0a6","Type":"ContainerDied","Data":"1b67b47da4a52585a88d6873c6a6f7a8bb324cb24c8579f5bb9a0fc23cbaf42e"} Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.264331 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.268427 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvwkb\" (UniqueName: \"kubernetes.io/projected/9511d654-cff2-45ce-9b6a-a3d98744e0a6-kube-api-access-dvwkb\") pod \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.268732 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-inventory\") pod \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.268971 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-bootstrap-combined-ca-bundle\") pod \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.269051 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-ssh-key\") pod \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\" (UID: \"9511d654-cff2-45ce-9b6a-a3d98744e0a6\") " Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.284558 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "9511d654-cff2-45ce-9b6a-a3d98744e0a6" (UID: "9511d654-cff2-45ce-9b6a-a3d98744e0a6"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.294604 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9511d654-cff2-45ce-9b6a-a3d98744e0a6-kube-api-access-dvwkb" (OuterVolumeSpecName: "kube-api-access-dvwkb") pod "9511d654-cff2-45ce-9b6a-a3d98744e0a6" (UID: "9511d654-cff2-45ce-9b6a-a3d98744e0a6"). InnerVolumeSpecName "kube-api-access-dvwkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.326574 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-inventory" (OuterVolumeSpecName: "inventory") pod "9511d654-cff2-45ce-9b6a-a3d98744e0a6" (UID: "9511d654-cff2-45ce-9b6a-a3d98744e0a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.349509 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9511d654-cff2-45ce-9b6a-a3d98744e0a6" (UID: "9511d654-cff2-45ce-9b6a-a3d98744e0a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.373053 4744 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.373099 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.373113 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvwkb\" (UniqueName: \"kubernetes.io/projected/9511d654-cff2-45ce-9b6a-a3d98744e0a6-kube-api-access-dvwkb\") on node \"crc\" DevicePath \"\"" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.373126 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9511d654-cff2-45ce-9b6a-a3d98744e0a6-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.700861 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" event={"ID":"9511d654-cff2-45ce-9b6a-a3d98744e0a6","Type":"ContainerDied","Data":"3e0ee4acc03a6f0692e31795a6ffc471a9d125e33dccc60648b9cd2f8cc64b4b"} Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.700906 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e0ee4acc03a6f0692e31795a6ffc471a9d125e33dccc60648b9cd2f8cc64b4b" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.700992 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.851210 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh"] Oct 08 09:37:00 crc kubenswrapper[4744]: E1008 09:37:00.851774 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9511d654-cff2-45ce-9b6a-a3d98744e0a6" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.851795 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9511d654-cff2-45ce-9b6a-a3d98744e0a6" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 09:37:00 crc kubenswrapper[4744]: E1008 09:37:00.851818 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2c1c075-2b01-4ce4-b851-9d15ee9afa57" containerName="registry-server" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.851824 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2c1c075-2b01-4ce4-b851-9d15ee9afa57" containerName="registry-server" Oct 08 09:37:00 crc kubenswrapper[4744]: E1008 09:37:00.851852 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2c1c075-2b01-4ce4-b851-9d15ee9afa57" containerName="extract-utilities" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.851858 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2c1c075-2b01-4ce4-b851-9d15ee9afa57" containerName="extract-utilities" Oct 08 09:37:00 crc kubenswrapper[4744]: E1008 09:37:00.851878 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2c1c075-2b01-4ce4-b851-9d15ee9afa57" containerName="extract-content" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.851883 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2c1c075-2b01-4ce4-b851-9d15ee9afa57" containerName="extract-content" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.852087 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2c1c075-2b01-4ce4-b851-9d15ee9afa57" containerName="registry-server" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.852105 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9511d654-cff2-45ce-9b6a-a3d98744e0a6" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.852945 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.856205 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.856482 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.856664 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.856789 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.879425 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh"] Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.985733 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-687kh\" (UID: \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.986092 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzz86\" (UniqueName: \"kubernetes.io/projected/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-kube-api-access-xzz86\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-687kh\" (UID: \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:37:00 crc kubenswrapper[4744]: I1008 09:37:00.986176 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-687kh\" (UID: \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:37:01 crc kubenswrapper[4744]: I1008 09:37:01.088789 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-687kh\" (UID: \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:37:01 crc kubenswrapper[4744]: I1008 09:37:01.089296 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-687kh\" (UID: \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:37:01 crc kubenswrapper[4744]: I1008 09:37:01.089480 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzz86\" (UniqueName: \"kubernetes.io/projected/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-kube-api-access-xzz86\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-687kh\" (UID: \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:37:01 crc kubenswrapper[4744]: I1008 09:37:01.096531 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-687kh\" (UID: \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:37:01 crc kubenswrapper[4744]: I1008 09:37:01.099205 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-687kh\" (UID: \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:37:01 crc kubenswrapper[4744]: I1008 09:37:01.109918 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzz86\" (UniqueName: \"kubernetes.io/projected/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-kube-api-access-xzz86\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-687kh\" (UID: \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:37:01 crc kubenswrapper[4744]: I1008 09:37:01.177004 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:37:01 crc kubenswrapper[4744]: I1008 09:37:01.828061 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 09:37:01 crc kubenswrapper[4744]: I1008 09:37:01.849247 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh"] Oct 08 09:37:02 crc kubenswrapper[4744]: I1008 09:37:02.746937 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" event={"ID":"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e","Type":"ContainerStarted","Data":"2ee85605b95bfb7b4c80f58f3e7d93f8c30f3a39fb7bbdd463384c8c50eea904"} Oct 08 09:37:02 crc kubenswrapper[4744]: I1008 09:37:02.747529 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" event={"ID":"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e","Type":"ContainerStarted","Data":"d6442a49c7f9bba4b2f4e32f36d803f53fcb8a9ddd3e13848f443bfd4297dba1"} Oct 08 09:37:02 crc kubenswrapper[4744]: I1008 09:37:02.793733 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" podStartSLOduration=2.276544512 podStartE2EDuration="2.793670892s" podCreationTimestamp="2025-10-08 09:37:00 +0000 UTC" firstStartedPulling="2025-10-08 09:37:01.827769127 +0000 UTC m=+1517.075414366" lastFinishedPulling="2025-10-08 09:37:02.344895507 +0000 UTC m=+1517.592540746" observedRunningTime="2025-10-08 09:37:02.778135046 +0000 UTC m=+1518.025780285" watchObservedRunningTime="2025-10-08 09:37:02.793670892 +0000 UTC m=+1518.041316131" Oct 08 09:37:18 crc kubenswrapper[4744]: I1008 09:37:18.048830 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-r4dmt"] Oct 08 09:37:18 crc kubenswrapper[4744]: I1008 09:37:18.060734 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-r4dmt"] Oct 08 09:37:19 crc kubenswrapper[4744]: I1008 09:37:19.465364 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d700c659-73db-4dc3-a2de-37e754f2775e" path="/var/lib/kubelet/pods/d700c659-73db-4dc3-a2de-37e754f2775e/volumes" Oct 08 09:37:19 crc kubenswrapper[4744]: I1008 09:37:19.689938 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:37:19 crc kubenswrapper[4744]: I1008 09:37:19.689994 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:37:22 crc kubenswrapper[4744]: I1008 09:37:22.043249 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-rdzjs"] Oct 08 09:37:22 crc kubenswrapper[4744]: I1008 09:37:22.054342 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-rdzjs"] Oct 08 09:37:23 crc kubenswrapper[4744]: I1008 09:37:23.051033 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-t9jv2"] Oct 08 09:37:23 crc kubenswrapper[4744]: I1008 09:37:23.063609 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-t9jv2"] Oct 08 09:37:23 crc kubenswrapper[4744]: I1008 09:37:23.465660 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d23fe19-d889-4aa8-9cc9-35eecb6449bc" path="/var/lib/kubelet/pods/1d23fe19-d889-4aa8-9cc9-35eecb6449bc/volumes" Oct 08 09:37:23 crc kubenswrapper[4744]: I1008 09:37:23.466536 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="744dc616-99e5-4b30-bd16-476de000e0ae" path="/var/lib/kubelet/pods/744dc616-99e5-4b30-bd16-476de000e0ae/volumes" Oct 08 09:37:24 crc kubenswrapper[4744]: I1008 09:37:24.792411 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x7zx6"] Oct 08 09:37:24 crc kubenswrapper[4744]: I1008 09:37:24.795499 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:24 crc kubenswrapper[4744]: I1008 09:37:24.830518 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x7zx6"] Oct 08 09:37:24 crc kubenswrapper[4744]: I1008 09:37:24.898478 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22d74b97-78a2-4acb-8cce-64b45599f882-catalog-content\") pod \"certified-operators-x7zx6\" (UID: \"22d74b97-78a2-4acb-8cce-64b45599f882\") " pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:24 crc kubenswrapper[4744]: I1008 09:37:24.898622 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22d74b97-78a2-4acb-8cce-64b45599f882-utilities\") pod \"certified-operators-x7zx6\" (UID: \"22d74b97-78a2-4acb-8cce-64b45599f882\") " pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:24 crc kubenswrapper[4744]: I1008 09:37:24.898667 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftpm2\" (UniqueName: \"kubernetes.io/projected/22d74b97-78a2-4acb-8cce-64b45599f882-kube-api-access-ftpm2\") pod \"certified-operators-x7zx6\" (UID: \"22d74b97-78a2-4acb-8cce-64b45599f882\") " pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:25 crc kubenswrapper[4744]: I1008 09:37:25.000804 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22d74b97-78a2-4acb-8cce-64b45599f882-utilities\") pod \"certified-operators-x7zx6\" (UID: \"22d74b97-78a2-4acb-8cce-64b45599f882\") " pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:25 crc kubenswrapper[4744]: I1008 09:37:25.000855 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftpm2\" (UniqueName: \"kubernetes.io/projected/22d74b97-78a2-4acb-8cce-64b45599f882-kube-api-access-ftpm2\") pod \"certified-operators-x7zx6\" (UID: \"22d74b97-78a2-4acb-8cce-64b45599f882\") " pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:25 crc kubenswrapper[4744]: I1008 09:37:25.000955 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22d74b97-78a2-4acb-8cce-64b45599f882-catalog-content\") pod \"certified-operators-x7zx6\" (UID: \"22d74b97-78a2-4acb-8cce-64b45599f882\") " pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:25 crc kubenswrapper[4744]: I1008 09:37:25.001328 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22d74b97-78a2-4acb-8cce-64b45599f882-catalog-content\") pod \"certified-operators-x7zx6\" (UID: \"22d74b97-78a2-4acb-8cce-64b45599f882\") " pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:25 crc kubenswrapper[4744]: I1008 09:37:25.001329 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22d74b97-78a2-4acb-8cce-64b45599f882-utilities\") pod \"certified-operators-x7zx6\" (UID: \"22d74b97-78a2-4acb-8cce-64b45599f882\") " pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:25 crc kubenswrapper[4744]: I1008 09:37:25.023050 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftpm2\" (UniqueName: \"kubernetes.io/projected/22d74b97-78a2-4acb-8cce-64b45599f882-kube-api-access-ftpm2\") pod \"certified-operators-x7zx6\" (UID: \"22d74b97-78a2-4acb-8cce-64b45599f882\") " pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:25 crc kubenswrapper[4744]: I1008 09:37:25.169020 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:25 crc kubenswrapper[4744]: I1008 09:37:25.667474 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x7zx6"] Oct 08 09:37:25 crc kubenswrapper[4744]: I1008 09:37:25.983132 4744 generic.go:334] "Generic (PLEG): container finished" podID="22d74b97-78a2-4acb-8cce-64b45599f882" containerID="64f2f4f643e97f956d3c4b820dfaaee7322c837bc702214ffe528b0922201172" exitCode=0 Oct 08 09:37:25 crc kubenswrapper[4744]: I1008 09:37:25.983183 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7zx6" event={"ID":"22d74b97-78a2-4acb-8cce-64b45599f882","Type":"ContainerDied","Data":"64f2f4f643e97f956d3c4b820dfaaee7322c837bc702214ffe528b0922201172"} Oct 08 09:37:25 crc kubenswrapper[4744]: I1008 09:37:25.983234 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7zx6" event={"ID":"22d74b97-78a2-4acb-8cce-64b45599f882","Type":"ContainerStarted","Data":"cf3273d1e842e684ebc6d45e68976f90d4ff7442b5ef87b2ebf8d912a477cb78"} Oct 08 09:37:28 crc kubenswrapper[4744]: I1008 09:37:28.007154 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7zx6" event={"ID":"22d74b97-78a2-4acb-8cce-64b45599f882","Type":"ContainerStarted","Data":"0e1615a03f58feb6ac290b13d60fcd769a9c452cbde1766dd11c0cd34cc458ad"} Oct 08 09:37:28 crc kubenswrapper[4744]: I1008 09:37:28.049687 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b2c5-account-create-lhn7c"] Oct 08 09:37:28 crc kubenswrapper[4744]: I1008 09:37:28.061211 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b2c5-account-create-lhn7c"] Oct 08 09:37:29 crc kubenswrapper[4744]: I1008 09:37:29.028665 4744 generic.go:334] "Generic (PLEG): container finished" podID="22d74b97-78a2-4acb-8cce-64b45599f882" containerID="0e1615a03f58feb6ac290b13d60fcd769a9c452cbde1766dd11c0cd34cc458ad" exitCode=0 Oct 08 09:37:29 crc kubenswrapper[4744]: I1008 09:37:29.028754 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7zx6" event={"ID":"22d74b97-78a2-4acb-8cce-64b45599f882","Type":"ContainerDied","Data":"0e1615a03f58feb6ac290b13d60fcd769a9c452cbde1766dd11c0cd34cc458ad"} Oct 08 09:37:29 crc kubenswrapper[4744]: I1008 09:37:29.463354 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00f05c86-12ae-4a9d-bc7c-8c4619a68028" path="/var/lib/kubelet/pods/00f05c86-12ae-4a9d-bc7c-8c4619a68028/volumes" Oct 08 09:37:30 crc kubenswrapper[4744]: I1008 09:37:30.043404 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7zx6" event={"ID":"22d74b97-78a2-4acb-8cce-64b45599f882","Type":"ContainerStarted","Data":"96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10"} Oct 08 09:37:30 crc kubenswrapper[4744]: I1008 09:37:30.075234 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x7zx6" podStartSLOduration=2.50535879 podStartE2EDuration="6.075215496s" podCreationTimestamp="2025-10-08 09:37:24 +0000 UTC" firstStartedPulling="2025-10-08 09:37:25.985387924 +0000 UTC m=+1541.233033163" lastFinishedPulling="2025-10-08 09:37:29.55524464 +0000 UTC m=+1544.802889869" observedRunningTime="2025-10-08 09:37:30.068354332 +0000 UTC m=+1545.315999571" watchObservedRunningTime="2025-10-08 09:37:30.075215496 +0000 UTC m=+1545.322860735" Oct 08 09:37:32 crc kubenswrapper[4744]: I1008 09:37:32.027228 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-a44f-account-create-qsjwh"] Oct 08 09:37:32 crc kubenswrapper[4744]: I1008 09:37:32.040541 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-a44f-account-create-qsjwh"] Oct 08 09:37:33 crc kubenswrapper[4744]: I1008 09:37:33.058819 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-f3a6-account-create-2cnwz"] Oct 08 09:37:33 crc kubenswrapper[4744]: I1008 09:37:33.071079 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-f3a6-account-create-2cnwz"] Oct 08 09:37:33 crc kubenswrapper[4744]: I1008 09:37:33.464362 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39dc9dd6-c3f6-4f28-b66a-3908b2585c5d" path="/var/lib/kubelet/pods/39dc9dd6-c3f6-4f28-b66a-3908b2585c5d/volumes" Oct 08 09:37:33 crc kubenswrapper[4744]: I1008 09:37:33.465298 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5b99de0-ebdd-460b-a6e9-011911768480" path="/var/lib/kubelet/pods/b5b99de0-ebdd-460b-a6e9-011911768480/volumes" Oct 08 09:37:35 crc kubenswrapper[4744]: I1008 09:37:35.169860 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:35 crc kubenswrapper[4744]: I1008 09:37:35.170230 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:35 crc kubenswrapper[4744]: I1008 09:37:35.218199 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:36 crc kubenswrapper[4744]: I1008 09:37:36.172584 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:36 crc kubenswrapper[4744]: I1008 09:37:36.237011 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x7zx6"] Oct 08 09:37:38 crc kubenswrapper[4744]: I1008 09:37:38.119412 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x7zx6" podUID="22d74b97-78a2-4acb-8cce-64b45599f882" containerName="registry-server" containerID="cri-o://96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10" gracePeriod=2 Oct 08 09:37:38 crc kubenswrapper[4744]: I1008 09:37:38.621412 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:38 crc kubenswrapper[4744]: I1008 09:37:38.691118 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22d74b97-78a2-4acb-8cce-64b45599f882-utilities\") pod \"22d74b97-78a2-4acb-8cce-64b45599f882\" (UID: \"22d74b97-78a2-4acb-8cce-64b45599f882\") " Oct 08 09:37:38 crc kubenswrapper[4744]: I1008 09:37:38.691764 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22d74b97-78a2-4acb-8cce-64b45599f882-catalog-content\") pod \"22d74b97-78a2-4acb-8cce-64b45599f882\" (UID: \"22d74b97-78a2-4acb-8cce-64b45599f882\") " Oct 08 09:37:38 crc kubenswrapper[4744]: I1008 09:37:38.692006 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftpm2\" (UniqueName: \"kubernetes.io/projected/22d74b97-78a2-4acb-8cce-64b45599f882-kube-api-access-ftpm2\") pod \"22d74b97-78a2-4acb-8cce-64b45599f882\" (UID: \"22d74b97-78a2-4acb-8cce-64b45599f882\") " Oct 08 09:37:38 crc kubenswrapper[4744]: I1008 09:37:38.692271 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22d74b97-78a2-4acb-8cce-64b45599f882-utilities" (OuterVolumeSpecName: "utilities") pod "22d74b97-78a2-4acb-8cce-64b45599f882" (UID: "22d74b97-78a2-4acb-8cce-64b45599f882"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:37:38 crc kubenswrapper[4744]: I1008 09:37:38.694595 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22d74b97-78a2-4acb-8cce-64b45599f882-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:37:38 crc kubenswrapper[4744]: I1008 09:37:38.710413 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22d74b97-78a2-4acb-8cce-64b45599f882-kube-api-access-ftpm2" (OuterVolumeSpecName: "kube-api-access-ftpm2") pod "22d74b97-78a2-4acb-8cce-64b45599f882" (UID: "22d74b97-78a2-4acb-8cce-64b45599f882"). InnerVolumeSpecName "kube-api-access-ftpm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:37:38 crc kubenswrapper[4744]: I1008 09:37:38.753289 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22d74b97-78a2-4acb-8cce-64b45599f882-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22d74b97-78a2-4acb-8cce-64b45599f882" (UID: "22d74b97-78a2-4acb-8cce-64b45599f882"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:37:38 crc kubenswrapper[4744]: I1008 09:37:38.796402 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22d74b97-78a2-4acb-8cce-64b45599f882-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:37:38 crc kubenswrapper[4744]: I1008 09:37:38.796612 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftpm2\" (UniqueName: \"kubernetes.io/projected/22d74b97-78a2-4acb-8cce-64b45599f882-kube-api-access-ftpm2\") on node \"crc\" DevicePath \"\"" Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.135272 4744 generic.go:334] "Generic (PLEG): container finished" podID="22d74b97-78a2-4acb-8cce-64b45599f882" containerID="96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10" exitCode=0 Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.135336 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x7zx6" Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.135342 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7zx6" event={"ID":"22d74b97-78a2-4acb-8cce-64b45599f882","Type":"ContainerDied","Data":"96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10"} Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.135601 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7zx6" event={"ID":"22d74b97-78a2-4acb-8cce-64b45599f882","Type":"ContainerDied","Data":"cf3273d1e842e684ebc6d45e68976f90d4ff7442b5ef87b2ebf8d912a477cb78"} Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.135633 4744 scope.go:117] "RemoveContainer" containerID="96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10" Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.158348 4744 scope.go:117] "RemoveContainer" containerID="0e1615a03f58feb6ac290b13d60fcd769a9c452cbde1766dd11c0cd34cc458ad" Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.176747 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x7zx6"] Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.187805 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x7zx6"] Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.195169 4744 scope.go:117] "RemoveContainer" containerID="64f2f4f643e97f956d3c4b820dfaaee7322c837bc702214ffe528b0922201172" Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.230901 4744 scope.go:117] "RemoveContainer" containerID="96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10" Oct 08 09:37:39 crc kubenswrapper[4744]: E1008 09:37:39.231643 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10\": container with ID starting with 96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10 not found: ID does not exist" containerID="96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10" Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.231682 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10"} err="failed to get container status \"96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10\": rpc error: code = NotFound desc = could not find container \"96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10\": container with ID starting with 96d2c1e8b7d3db3aaf6e737cdbec014c01d990b12934b2593d71d64586b0bf10 not found: ID does not exist" Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.231710 4744 scope.go:117] "RemoveContainer" containerID="0e1615a03f58feb6ac290b13d60fcd769a9c452cbde1766dd11c0cd34cc458ad" Oct 08 09:37:39 crc kubenswrapper[4744]: E1008 09:37:39.232700 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e1615a03f58feb6ac290b13d60fcd769a9c452cbde1766dd11c0cd34cc458ad\": container with ID starting with 0e1615a03f58feb6ac290b13d60fcd769a9c452cbde1766dd11c0cd34cc458ad not found: ID does not exist" containerID="0e1615a03f58feb6ac290b13d60fcd769a9c452cbde1766dd11c0cd34cc458ad" Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.232729 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e1615a03f58feb6ac290b13d60fcd769a9c452cbde1766dd11c0cd34cc458ad"} err="failed to get container status \"0e1615a03f58feb6ac290b13d60fcd769a9c452cbde1766dd11c0cd34cc458ad\": rpc error: code = NotFound desc = could not find container \"0e1615a03f58feb6ac290b13d60fcd769a9c452cbde1766dd11c0cd34cc458ad\": container with ID starting with 0e1615a03f58feb6ac290b13d60fcd769a9c452cbde1766dd11c0cd34cc458ad not found: ID does not exist" Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.232751 4744 scope.go:117] "RemoveContainer" containerID="64f2f4f643e97f956d3c4b820dfaaee7322c837bc702214ffe528b0922201172" Oct 08 09:37:39 crc kubenswrapper[4744]: E1008 09:37:39.233060 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64f2f4f643e97f956d3c4b820dfaaee7322c837bc702214ffe528b0922201172\": container with ID starting with 64f2f4f643e97f956d3c4b820dfaaee7322c837bc702214ffe528b0922201172 not found: ID does not exist" containerID="64f2f4f643e97f956d3c4b820dfaaee7322c837bc702214ffe528b0922201172" Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.233082 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64f2f4f643e97f956d3c4b820dfaaee7322c837bc702214ffe528b0922201172"} err="failed to get container status \"64f2f4f643e97f956d3c4b820dfaaee7322c837bc702214ffe528b0922201172\": rpc error: code = NotFound desc = could not find container \"64f2f4f643e97f956d3c4b820dfaaee7322c837bc702214ffe528b0922201172\": container with ID starting with 64f2f4f643e97f956d3c4b820dfaaee7322c837bc702214ffe528b0922201172 not found: ID does not exist" Oct 08 09:37:39 crc kubenswrapper[4744]: I1008 09:37:39.466331 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22d74b97-78a2-4acb-8cce-64b45599f882" path="/var/lib/kubelet/pods/22d74b97-78a2-4acb-8cce-64b45599f882/volumes" Oct 08 09:37:49 crc kubenswrapper[4744]: I1008 09:37:49.690287 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:37:49 crc kubenswrapper[4744]: I1008 09:37:49.690872 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:37:49 crc kubenswrapper[4744]: I1008 09:37:49.690923 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:37:49 crc kubenswrapper[4744]: I1008 09:37:49.691728 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 09:37:49 crc kubenswrapper[4744]: I1008 09:37:49.691798 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" gracePeriod=600 Oct 08 09:37:49 crc kubenswrapper[4744]: E1008 09:37:49.815780 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:37:50 crc kubenswrapper[4744]: I1008 09:37:50.263689 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" exitCode=0 Oct 08 09:37:50 crc kubenswrapper[4744]: I1008 09:37:50.263899 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb"} Oct 08 09:37:50 crc kubenswrapper[4744]: I1008 09:37:50.264084 4744 scope.go:117] "RemoveContainer" containerID="936d206815e6d287935cf3c047d48aada7c4fce45edc4c83541f33b9de7f8702" Oct 08 09:37:50 crc kubenswrapper[4744]: I1008 09:37:50.264878 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:37:50 crc kubenswrapper[4744]: E1008 09:37:50.265351 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:37:53 crc kubenswrapper[4744]: I1008 09:37:53.247045 4744 scope.go:117] "RemoveContainer" containerID="c76c311a1bafdd7b9488148f76accf280f866bf3c9bc6f06367c38f3dbf50b6b" Oct 08 09:37:53 crc kubenswrapper[4744]: I1008 09:37:53.270327 4744 scope.go:117] "RemoveContainer" containerID="cffa31b9a562fe8c7bfb9599d87bf1723bf56195baef526033907924db285758" Oct 08 09:37:53 crc kubenswrapper[4744]: I1008 09:37:53.312044 4744 scope.go:117] "RemoveContainer" containerID="3437fcc9c340f5c61a710ed66d4b69c51113ba5cd0bcd4197aa2b7ed20ba27e7" Oct 08 09:37:53 crc kubenswrapper[4744]: I1008 09:37:53.337067 4744 scope.go:117] "RemoveContainer" containerID="8b9cf5919c6fc5f275d0718804ff1e4bc1028d06cbd99118545f053b45af183c" Oct 08 09:37:53 crc kubenswrapper[4744]: I1008 09:37:53.370147 4744 scope.go:117] "RemoveContainer" containerID="e65f207bcf2f14ccb82b865966626dc284f7f15e57faebc3d79703a496df6b10" Oct 08 09:37:53 crc kubenswrapper[4744]: I1008 09:37:53.414247 4744 scope.go:117] "RemoveContainer" containerID="ea33cb02115338e3fabbb430fe5159429bf7ce934db51851c504a0c1ceca3574" Oct 08 09:37:53 crc kubenswrapper[4744]: I1008 09:37:53.462357 4744 scope.go:117] "RemoveContainer" containerID="977cd4c09fdbc0d482979d07803fa655b3235316582e09225a68e1e2e243e9cb" Oct 08 09:37:53 crc kubenswrapper[4744]: I1008 09:37:53.503109 4744 scope.go:117] "RemoveContainer" containerID="bebda48be52e4d4862462d1a9b8d02e52be152f93dd5281c01b88bec75eec8c4" Oct 08 09:37:56 crc kubenswrapper[4744]: I1008 09:37:56.035452 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-fdrfp"] Oct 08 09:37:56 crc kubenswrapper[4744]: I1008 09:37:56.045871 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-fdrfp"] Oct 08 09:37:57 crc kubenswrapper[4744]: I1008 09:37:57.463762 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f33787b3-d366-41b9-80f7-955dc458cb93" path="/var/lib/kubelet/pods/f33787b3-d366-41b9-80f7-955dc458cb93/volumes" Oct 08 09:38:04 crc kubenswrapper[4744]: I1008 09:38:04.453641 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:38:04 crc kubenswrapper[4744]: E1008 09:38:04.454413 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:38:11 crc kubenswrapper[4744]: I1008 09:38:11.064210 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-7rjbs"] Oct 08 09:38:11 crc kubenswrapper[4744]: I1008 09:38:11.078848 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-7rjbs"] Oct 08 09:38:11 crc kubenswrapper[4744]: I1008 09:38:11.095337 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-g7pft"] Oct 08 09:38:11 crc kubenswrapper[4744]: I1008 09:38:11.108350 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-pv2cr"] Oct 08 09:38:11 crc kubenswrapper[4744]: I1008 09:38:11.120179 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-g7pft"] Oct 08 09:38:11 crc kubenswrapper[4744]: I1008 09:38:11.130651 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-pv2cr"] Oct 08 09:38:11 crc kubenswrapper[4744]: I1008 09:38:11.464700 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="296ff709-3fab-4f25-ba36-7476d1a2b9d1" path="/var/lib/kubelet/pods/296ff709-3fab-4f25-ba36-7476d1a2b9d1/volumes" Oct 08 09:38:11 crc kubenswrapper[4744]: I1008 09:38:11.465928 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73553460-0850-41da-8ae1-33ced6d028f7" path="/var/lib/kubelet/pods/73553460-0850-41da-8ae1-33ced6d028f7/volumes" Oct 08 09:38:11 crc kubenswrapper[4744]: I1008 09:38:11.466651 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7706a783-6b39-431f-b765-e189122a38e2" path="/var/lib/kubelet/pods/7706a783-6b39-431f-b765-e189122a38e2/volumes" Oct 08 09:38:17 crc kubenswrapper[4744]: I1008 09:38:17.453716 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:38:17 crc kubenswrapper[4744]: E1008 09:38:17.454641 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:38:20 crc kubenswrapper[4744]: I1008 09:38:20.063460 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c3ca-account-create-nftjn"] Oct 08 09:38:20 crc kubenswrapper[4744]: I1008 09:38:20.078734 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c3ca-account-create-nftjn"] Oct 08 09:38:21 crc kubenswrapper[4744]: I1008 09:38:21.035497 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-zhx7v"] Oct 08 09:38:21 crc kubenswrapper[4744]: I1008 09:38:21.045730 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-zhx7v"] Oct 08 09:38:21 crc kubenswrapper[4744]: I1008 09:38:21.465680 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a658cb27-4e1d-4d3b-90cd-1e16b4013c48" path="/var/lib/kubelet/pods/a658cb27-4e1d-4d3b-90cd-1e16b4013c48/volumes" Oct 08 09:38:21 crc kubenswrapper[4744]: I1008 09:38:21.466488 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bab2aef1-a444-4b40-bcc3-b59ba4755bf7" path="/var/lib/kubelet/pods/bab2aef1-a444-4b40-bcc3-b59ba4755bf7/volumes" Oct 08 09:38:22 crc kubenswrapper[4744]: I1008 09:38:22.615028 4744 generic.go:334] "Generic (PLEG): container finished" podID="c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e" containerID="2ee85605b95bfb7b4c80f58f3e7d93f8c30f3a39fb7bbdd463384c8c50eea904" exitCode=0 Oct 08 09:38:22 crc kubenswrapper[4744]: I1008 09:38:22.615117 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" event={"ID":"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e","Type":"ContainerDied","Data":"2ee85605b95bfb7b4c80f58f3e7d93f8c30f3a39fb7bbdd463384c8c50eea904"} Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.135094 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.178640 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-inventory\") pod \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\" (UID: \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\") " Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.178762 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-ssh-key\") pod \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\" (UID: \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\") " Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.178874 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzz86\" (UniqueName: \"kubernetes.io/projected/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-kube-api-access-xzz86\") pod \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\" (UID: \"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e\") " Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.187342 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-kube-api-access-xzz86" (OuterVolumeSpecName: "kube-api-access-xzz86") pod "c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e" (UID: "c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e"). InnerVolumeSpecName "kube-api-access-xzz86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.215586 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e" (UID: "c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.219560 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-inventory" (OuterVolumeSpecName: "inventory") pod "c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e" (UID: "c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.280916 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.280984 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzz86\" (UniqueName: \"kubernetes.io/projected/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-kube-api-access-xzz86\") on node \"crc\" DevicePath \"\"" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.281027 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.640091 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" event={"ID":"c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e","Type":"ContainerDied","Data":"d6442a49c7f9bba4b2f4e32f36d803f53fcb8a9ddd3e13848f443bfd4297dba1"} Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.640525 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6442a49c7f9bba4b2f4e32f36d803f53fcb8a9ddd3e13848f443bfd4297dba1" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.640208 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.743728 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms"] Oct 08 09:38:24 crc kubenswrapper[4744]: E1008 09:38:24.747258 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.747303 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 09:38:24 crc kubenswrapper[4744]: E1008 09:38:24.747321 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22d74b97-78a2-4acb-8cce-64b45599f882" containerName="extract-content" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.747332 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="22d74b97-78a2-4acb-8cce-64b45599f882" containerName="extract-content" Oct 08 09:38:24 crc kubenswrapper[4744]: E1008 09:38:24.747630 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22d74b97-78a2-4acb-8cce-64b45599f882" containerName="extract-utilities" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.747652 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="22d74b97-78a2-4acb-8cce-64b45599f882" containerName="extract-utilities" Oct 08 09:38:24 crc kubenswrapper[4744]: E1008 09:38:24.747676 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22d74b97-78a2-4acb-8cce-64b45599f882" containerName="registry-server" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.747685 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="22d74b97-78a2-4acb-8cce-64b45599f882" containerName="registry-server" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.748038 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="22d74b97-78a2-4acb-8cce-64b45599f882" containerName="registry-server" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.748059 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.749078 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.757300 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.757592 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.757783 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.758213 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.790273 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ae696f3-08f6-4219-8b34-eb9afe223239-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms\" (UID: \"7ae696f3-08f6-4219-8b34-eb9afe223239\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.790577 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t99j\" (UniqueName: \"kubernetes.io/projected/7ae696f3-08f6-4219-8b34-eb9afe223239-kube-api-access-6t99j\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms\" (UID: \"7ae696f3-08f6-4219-8b34-eb9afe223239\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.790762 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ae696f3-08f6-4219-8b34-eb9afe223239-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms\" (UID: \"7ae696f3-08f6-4219-8b34-eb9afe223239\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.794950 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms"] Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.892180 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ae696f3-08f6-4219-8b34-eb9afe223239-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms\" (UID: \"7ae696f3-08f6-4219-8b34-eb9afe223239\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.892767 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ae696f3-08f6-4219-8b34-eb9afe223239-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms\" (UID: \"7ae696f3-08f6-4219-8b34-eb9afe223239\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.892793 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t99j\" (UniqueName: \"kubernetes.io/projected/7ae696f3-08f6-4219-8b34-eb9afe223239-kube-api-access-6t99j\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms\" (UID: \"7ae696f3-08f6-4219-8b34-eb9afe223239\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.900268 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ae696f3-08f6-4219-8b34-eb9afe223239-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms\" (UID: \"7ae696f3-08f6-4219-8b34-eb9afe223239\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.900304 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ae696f3-08f6-4219-8b34-eb9afe223239-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms\" (UID: \"7ae696f3-08f6-4219-8b34-eb9afe223239\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:24 crc kubenswrapper[4744]: I1008 09:38:24.915193 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t99j\" (UniqueName: \"kubernetes.io/projected/7ae696f3-08f6-4219-8b34-eb9afe223239-kube-api-access-6t99j\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms\" (UID: \"7ae696f3-08f6-4219-8b34-eb9afe223239\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:25 crc kubenswrapper[4744]: I1008 09:38:25.096357 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:25 crc kubenswrapper[4744]: I1008 09:38:25.679076 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms"] Oct 08 09:38:26 crc kubenswrapper[4744]: I1008 09:38:26.665660 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" event={"ID":"7ae696f3-08f6-4219-8b34-eb9afe223239","Type":"ContainerStarted","Data":"9e07c94b16ead85a8580ea2841cf7884b57509e2d083a5ab1bb80ff46a16c36a"} Oct 08 09:38:26 crc kubenswrapper[4744]: I1008 09:38:26.666181 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" event={"ID":"7ae696f3-08f6-4219-8b34-eb9afe223239","Type":"ContainerStarted","Data":"908cf6d5f08ada795dd52ad567381a24c4c8982689702fe2e9daa3c40ec220de"} Oct 08 09:38:26 crc kubenswrapper[4744]: I1008 09:38:26.700088 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" podStartSLOduration=2.192663065 podStartE2EDuration="2.700054419s" podCreationTimestamp="2025-10-08 09:38:24 +0000 UTC" firstStartedPulling="2025-10-08 09:38:25.685069695 +0000 UTC m=+1600.932714934" lastFinishedPulling="2025-10-08 09:38:26.192461049 +0000 UTC m=+1601.440106288" observedRunningTime="2025-10-08 09:38:26.693216067 +0000 UTC m=+1601.940861316" watchObservedRunningTime="2025-10-08 09:38:26.700054419 +0000 UTC m=+1601.947699668" Oct 08 09:38:31 crc kubenswrapper[4744]: I1008 09:38:31.722480 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ae696f3-08f6-4219-8b34-eb9afe223239" containerID="9e07c94b16ead85a8580ea2841cf7884b57509e2d083a5ab1bb80ff46a16c36a" exitCode=0 Oct 08 09:38:31 crc kubenswrapper[4744]: I1008 09:38:31.722561 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" event={"ID":"7ae696f3-08f6-4219-8b34-eb9afe223239","Type":"ContainerDied","Data":"9e07c94b16ead85a8580ea2841cf7884b57509e2d083a5ab1bb80ff46a16c36a"} Oct 08 09:38:32 crc kubenswrapper[4744]: I1008 09:38:32.453210 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:38:32 crc kubenswrapper[4744]: E1008 09:38:32.454069 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.076037 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-86e2-account-create-7jnn7"] Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.089644 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-86e2-account-create-7jnn7"] Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.211547 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.290746 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t99j\" (UniqueName: \"kubernetes.io/projected/7ae696f3-08f6-4219-8b34-eb9afe223239-kube-api-access-6t99j\") pod \"7ae696f3-08f6-4219-8b34-eb9afe223239\" (UID: \"7ae696f3-08f6-4219-8b34-eb9afe223239\") " Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.290840 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ae696f3-08f6-4219-8b34-eb9afe223239-inventory\") pod \"7ae696f3-08f6-4219-8b34-eb9afe223239\" (UID: \"7ae696f3-08f6-4219-8b34-eb9afe223239\") " Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.291010 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ae696f3-08f6-4219-8b34-eb9afe223239-ssh-key\") pod \"7ae696f3-08f6-4219-8b34-eb9afe223239\" (UID: \"7ae696f3-08f6-4219-8b34-eb9afe223239\") " Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.298025 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ae696f3-08f6-4219-8b34-eb9afe223239-kube-api-access-6t99j" (OuterVolumeSpecName: "kube-api-access-6t99j") pod "7ae696f3-08f6-4219-8b34-eb9afe223239" (UID: "7ae696f3-08f6-4219-8b34-eb9afe223239"). InnerVolumeSpecName "kube-api-access-6t99j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.319109 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ae696f3-08f6-4219-8b34-eb9afe223239-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7ae696f3-08f6-4219-8b34-eb9afe223239" (UID: "7ae696f3-08f6-4219-8b34-eb9afe223239"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.324549 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ae696f3-08f6-4219-8b34-eb9afe223239-inventory" (OuterVolumeSpecName: "inventory") pod "7ae696f3-08f6-4219-8b34-eb9afe223239" (UID: "7ae696f3-08f6-4219-8b34-eb9afe223239"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.392473 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ae696f3-08f6-4219-8b34-eb9afe223239-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.392879 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t99j\" (UniqueName: \"kubernetes.io/projected/7ae696f3-08f6-4219-8b34-eb9afe223239-kube-api-access-6t99j\") on node \"crc\" DevicePath \"\"" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.392891 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ae696f3-08f6-4219-8b34-eb9afe223239-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.464757 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1cf3771-9139-48f3-9d3b-9533b957aa14" path="/var/lib/kubelet/pods/d1cf3771-9139-48f3-9d3b-9533b957aa14/volumes" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.746601 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" event={"ID":"7ae696f3-08f6-4219-8b34-eb9afe223239","Type":"ContainerDied","Data":"908cf6d5f08ada795dd52ad567381a24c4c8982689702fe2e9daa3c40ec220de"} Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.746674 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="908cf6d5f08ada795dd52ad567381a24c4c8982689702fe2e9daa3c40ec220de" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.746712 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.872431 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4"] Oct 08 09:38:33 crc kubenswrapper[4744]: E1008 09:38:33.872895 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ae696f3-08f6-4219-8b34-eb9afe223239" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.872912 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ae696f3-08f6-4219-8b34-eb9afe223239" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.873119 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ae696f3-08f6-4219-8b34-eb9afe223239" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.873862 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.882043 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.882272 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.882462 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.882575 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.893465 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4"] Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.902608 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f69541a-2daf-4e28-b1a1-1811c5476652-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mnhk4\" (UID: \"7f69541a-2daf-4e28-b1a1-1811c5476652\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.902732 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htwmc\" (UniqueName: \"kubernetes.io/projected/7f69541a-2daf-4e28-b1a1-1811c5476652-kube-api-access-htwmc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mnhk4\" (UID: \"7f69541a-2daf-4e28-b1a1-1811c5476652\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:38:33 crc kubenswrapper[4744]: I1008 09:38:33.902782 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f69541a-2daf-4e28-b1a1-1811c5476652-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mnhk4\" (UID: \"7f69541a-2daf-4e28-b1a1-1811c5476652\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:38:34 crc kubenswrapper[4744]: I1008 09:38:34.004411 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f69541a-2daf-4e28-b1a1-1811c5476652-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mnhk4\" (UID: \"7f69541a-2daf-4e28-b1a1-1811c5476652\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:38:34 crc kubenswrapper[4744]: I1008 09:38:34.004542 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htwmc\" (UniqueName: \"kubernetes.io/projected/7f69541a-2daf-4e28-b1a1-1811c5476652-kube-api-access-htwmc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mnhk4\" (UID: \"7f69541a-2daf-4e28-b1a1-1811c5476652\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:38:34 crc kubenswrapper[4744]: I1008 09:38:34.004599 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f69541a-2daf-4e28-b1a1-1811c5476652-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mnhk4\" (UID: \"7f69541a-2daf-4e28-b1a1-1811c5476652\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:38:34 crc kubenswrapper[4744]: I1008 09:38:34.012919 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f69541a-2daf-4e28-b1a1-1811c5476652-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mnhk4\" (UID: \"7f69541a-2daf-4e28-b1a1-1811c5476652\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:38:34 crc kubenswrapper[4744]: I1008 09:38:34.016659 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f69541a-2daf-4e28-b1a1-1811c5476652-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mnhk4\" (UID: \"7f69541a-2daf-4e28-b1a1-1811c5476652\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:38:34 crc kubenswrapper[4744]: I1008 09:38:34.022759 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htwmc\" (UniqueName: \"kubernetes.io/projected/7f69541a-2daf-4e28-b1a1-1811c5476652-kube-api-access-htwmc\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mnhk4\" (UID: \"7f69541a-2daf-4e28-b1a1-1811c5476652\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:38:34 crc kubenswrapper[4744]: I1008 09:38:34.195569 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:38:34 crc kubenswrapper[4744]: I1008 09:38:34.781566 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4"] Oct 08 09:38:35 crc kubenswrapper[4744]: I1008 09:38:35.776320 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" event={"ID":"7f69541a-2daf-4e28-b1a1-1811c5476652","Type":"ContainerStarted","Data":"34274e263382b096e9cec1e24525e4ef27c4003bad261a860a35e16df357fd63"} Oct 08 09:38:35 crc kubenswrapper[4744]: I1008 09:38:35.776856 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" event={"ID":"7f69541a-2daf-4e28-b1a1-1811c5476652","Type":"ContainerStarted","Data":"a9cf687a3cb6cd5cf86ae6efa82d531a12b15ecdb65991c5cfca485c13c70b70"} Oct 08 09:38:35 crc kubenswrapper[4744]: I1008 09:38:35.805107 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" podStartSLOduration=2.333697741 podStartE2EDuration="2.805084825s" podCreationTimestamp="2025-10-08 09:38:33 +0000 UTC" firstStartedPulling="2025-10-08 09:38:34.81822132 +0000 UTC m=+1610.065866549" lastFinishedPulling="2025-10-08 09:38:35.289608394 +0000 UTC m=+1610.537253633" observedRunningTime="2025-10-08 09:38:35.794178379 +0000 UTC m=+1611.041823618" watchObservedRunningTime="2025-10-08 09:38:35.805084825 +0000 UTC m=+1611.052730064" Oct 08 09:38:36 crc kubenswrapper[4744]: I1008 09:38:36.042564 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-96af-account-create-47426"] Oct 08 09:38:36 crc kubenswrapper[4744]: I1008 09:38:36.053058 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-96af-account-create-47426"] Oct 08 09:38:37 crc kubenswrapper[4744]: I1008 09:38:37.464531 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af798bff-4019-44a0-a0f9-f6213e2c2d64" path="/var/lib/kubelet/pods/af798bff-4019-44a0-a0f9-f6213e2c2d64/volumes" Oct 08 09:38:38 crc kubenswrapper[4744]: I1008 09:38:38.045345 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-g7tp7"] Oct 08 09:38:38 crc kubenswrapper[4744]: I1008 09:38:38.052992 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-g7tp7"] Oct 08 09:38:39 crc kubenswrapper[4744]: I1008 09:38:39.484398 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6a227c0-ae15-427c-9557-72be160cb20c" path="/var/lib/kubelet/pods/a6a227c0-ae15-427c-9557-72be160cb20c/volumes" Oct 08 09:38:43 crc kubenswrapper[4744]: I1008 09:38:43.453899 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:38:43 crc kubenswrapper[4744]: E1008 09:38:43.454798 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:38:45 crc kubenswrapper[4744]: I1008 09:38:45.041399 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rgx4v"] Oct 08 09:38:45 crc kubenswrapper[4744]: I1008 09:38:45.055800 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rgx4v"] Oct 08 09:38:45 crc kubenswrapper[4744]: I1008 09:38:45.473105 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6aacf4c-187c-4d2c-a740-09514e340744" path="/var/lib/kubelet/pods/c6aacf4c-187c-4d2c-a740-09514e340744/volumes" Oct 08 09:38:53 crc kubenswrapper[4744]: I1008 09:38:53.042777 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-fxgnx"] Oct 08 09:38:53 crc kubenswrapper[4744]: I1008 09:38:53.052192 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-fxgnx"] Oct 08 09:38:53 crc kubenswrapper[4744]: I1008 09:38:53.471548 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54a05787-2fe8-4d6d-b785-058040260352" path="/var/lib/kubelet/pods/54a05787-2fe8-4d6d-b785-058040260352/volumes" Oct 08 09:38:53 crc kubenswrapper[4744]: I1008 09:38:53.654519 4744 scope.go:117] "RemoveContainer" containerID="b5e7b98c0e146c6e932e8602f8fcfd2d7d0c377be903294c753a386d73e111fb" Oct 08 09:38:53 crc kubenswrapper[4744]: I1008 09:38:53.710838 4744 scope.go:117] "RemoveContainer" containerID="5e8fb6af05a95d7d62e576edbfb946d591709c6c28b7dd46787bdf59e83da84d" Oct 08 09:38:53 crc kubenswrapper[4744]: I1008 09:38:53.746421 4744 scope.go:117] "RemoveContainer" containerID="2544b2efd54a6b3b29f077a1fd78405bc7739e74cf3844435e05fc27b230527e" Oct 08 09:38:53 crc kubenswrapper[4744]: I1008 09:38:53.836796 4744 scope.go:117] "RemoveContainer" containerID="60a038896f7a4eac5aceca441989984b484b422fb7dedc2d4488a50fb6b27c9f" Oct 08 09:38:53 crc kubenswrapper[4744]: I1008 09:38:53.886441 4744 scope.go:117] "RemoveContainer" containerID="b2a66635ddeb7399d57dd21ebf682ba5b1ca75fc8934a7145df8d3f597cffb92" Oct 08 09:38:53 crc kubenswrapper[4744]: I1008 09:38:53.917197 4744 scope.go:117] "RemoveContainer" containerID="6ef8d9d17f46150fccc932f1a7f082a31010389556aa962745653259a5f12e5d" Oct 08 09:38:53 crc kubenswrapper[4744]: I1008 09:38:53.976426 4744 scope.go:117] "RemoveContainer" containerID="559c32c64a550f3820d5cdcb0aff75cb41af4b0add66caaf9585bbc2d1a8587a" Oct 08 09:38:54 crc kubenswrapper[4744]: I1008 09:38:54.006582 4744 scope.go:117] "RemoveContainer" containerID="9dbaad1a6c92ea0a7e75066ec1bcb443c6516f24254fa7758c499382e3c376a3" Oct 08 09:38:54 crc kubenswrapper[4744]: I1008 09:38:54.031140 4744 scope.go:117] "RemoveContainer" containerID="ea4971b0bcfa3ea8271b5f449d7ac39af65ed01983276a03216c201f800f3d49" Oct 08 09:38:54 crc kubenswrapper[4744]: I1008 09:38:54.069976 4744 scope.go:117] "RemoveContainer" containerID="4e736fd721b1c7fc0f95275c73e3d9f42b2f078ba886f2fc7f4bc07720e721a4" Oct 08 09:38:54 crc kubenswrapper[4744]: I1008 09:38:54.093155 4744 scope.go:117] "RemoveContainer" containerID="c7a6b503518b670ce3301e8c0a603a1738d812039d0262a6307ad03d6cbbab63" Oct 08 09:38:56 crc kubenswrapper[4744]: I1008 09:38:56.453436 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:38:56 crc kubenswrapper[4744]: E1008 09:38:56.454146 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:39:00 crc kubenswrapper[4744]: I1008 09:39:00.063626 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-bxh5q"] Oct 08 09:39:00 crc kubenswrapper[4744]: I1008 09:39:00.070670 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-bxh5q"] Oct 08 09:39:01 crc kubenswrapper[4744]: I1008 09:39:01.473301 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f301473-6a0d-4b3d-936d-9078c6ea6c54" path="/var/lib/kubelet/pods/4f301473-6a0d-4b3d-936d-9078c6ea6c54/volumes" Oct 08 09:39:07 crc kubenswrapper[4744]: I1008 09:39:07.454110 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:39:07 crc kubenswrapper[4744]: E1008 09:39:07.455178 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:39:18 crc kubenswrapper[4744]: I1008 09:39:18.452986 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:39:18 crc kubenswrapper[4744]: E1008 09:39:18.453920 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:39:19 crc kubenswrapper[4744]: I1008 09:39:19.269033 4744 generic.go:334] "Generic (PLEG): container finished" podID="7f69541a-2daf-4e28-b1a1-1811c5476652" containerID="34274e263382b096e9cec1e24525e4ef27c4003bad261a860a35e16df357fd63" exitCode=0 Oct 08 09:39:19 crc kubenswrapper[4744]: I1008 09:39:19.269106 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" event={"ID":"7f69541a-2daf-4e28-b1a1-1811c5476652","Type":"ContainerDied","Data":"34274e263382b096e9cec1e24525e4ef27c4003bad261a860a35e16df357fd63"} Oct 08 09:39:20 crc kubenswrapper[4744]: I1008 09:39:20.773159 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:39:20 crc kubenswrapper[4744]: I1008 09:39:20.907416 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htwmc\" (UniqueName: \"kubernetes.io/projected/7f69541a-2daf-4e28-b1a1-1811c5476652-kube-api-access-htwmc\") pod \"7f69541a-2daf-4e28-b1a1-1811c5476652\" (UID: \"7f69541a-2daf-4e28-b1a1-1811c5476652\") " Oct 08 09:39:20 crc kubenswrapper[4744]: I1008 09:39:20.907793 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f69541a-2daf-4e28-b1a1-1811c5476652-ssh-key\") pod \"7f69541a-2daf-4e28-b1a1-1811c5476652\" (UID: \"7f69541a-2daf-4e28-b1a1-1811c5476652\") " Oct 08 09:39:20 crc kubenswrapper[4744]: I1008 09:39:20.907872 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f69541a-2daf-4e28-b1a1-1811c5476652-inventory\") pod \"7f69541a-2daf-4e28-b1a1-1811c5476652\" (UID: \"7f69541a-2daf-4e28-b1a1-1811c5476652\") " Oct 08 09:39:20 crc kubenswrapper[4744]: I1008 09:39:20.918149 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f69541a-2daf-4e28-b1a1-1811c5476652-kube-api-access-htwmc" (OuterVolumeSpecName: "kube-api-access-htwmc") pod "7f69541a-2daf-4e28-b1a1-1811c5476652" (UID: "7f69541a-2daf-4e28-b1a1-1811c5476652"). InnerVolumeSpecName "kube-api-access-htwmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:39:20 crc kubenswrapper[4744]: I1008 09:39:20.940885 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f69541a-2daf-4e28-b1a1-1811c5476652-inventory" (OuterVolumeSpecName: "inventory") pod "7f69541a-2daf-4e28-b1a1-1811c5476652" (UID: "7f69541a-2daf-4e28-b1a1-1811c5476652"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:39:20 crc kubenswrapper[4744]: I1008 09:39:20.948694 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f69541a-2daf-4e28-b1a1-1811c5476652-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7f69541a-2daf-4e28-b1a1-1811c5476652" (UID: "7f69541a-2daf-4e28-b1a1-1811c5476652"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.010347 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f69541a-2daf-4e28-b1a1-1811c5476652-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.010452 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f69541a-2daf-4e28-b1a1-1811c5476652-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.010465 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htwmc\" (UniqueName: \"kubernetes.io/projected/7f69541a-2daf-4e28-b1a1-1811c5476652-kube-api-access-htwmc\") on node \"crc\" DevicePath \"\"" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.293896 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" event={"ID":"7f69541a-2daf-4e28-b1a1-1811c5476652","Type":"ContainerDied","Data":"a9cf687a3cb6cd5cf86ae6efa82d531a12b15ecdb65991c5cfca485c13c70b70"} Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.293973 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9cf687a3cb6cd5cf86ae6efa82d531a12b15ecdb65991c5cfca485c13c70b70" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.294069 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.406094 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg"] Oct 08 09:39:21 crc kubenswrapper[4744]: E1008 09:39:21.406668 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f69541a-2daf-4e28-b1a1-1811c5476652" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.406691 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f69541a-2daf-4e28-b1a1-1811c5476652" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.406874 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f69541a-2daf-4e28-b1a1-1811c5476652" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.408275 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.412028 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.412276 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.412552 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.415721 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.418683 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg"] Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.520402 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26f17329-55b8-4de4-82ee-98775dafa32a-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg\" (UID: \"26f17329-55b8-4de4-82ee-98775dafa32a\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.520466 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26f17329-55b8-4de4-82ee-98775dafa32a-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg\" (UID: \"26f17329-55b8-4de4-82ee-98775dafa32a\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.520521 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzpfs\" (UniqueName: \"kubernetes.io/projected/26f17329-55b8-4de4-82ee-98775dafa32a-kube-api-access-pzpfs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg\" (UID: \"26f17329-55b8-4de4-82ee-98775dafa32a\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.622705 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26f17329-55b8-4de4-82ee-98775dafa32a-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg\" (UID: \"26f17329-55b8-4de4-82ee-98775dafa32a\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.622757 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26f17329-55b8-4de4-82ee-98775dafa32a-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg\" (UID: \"26f17329-55b8-4de4-82ee-98775dafa32a\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.622819 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzpfs\" (UniqueName: \"kubernetes.io/projected/26f17329-55b8-4de4-82ee-98775dafa32a-kube-api-access-pzpfs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg\" (UID: \"26f17329-55b8-4de4-82ee-98775dafa32a\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.627484 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26f17329-55b8-4de4-82ee-98775dafa32a-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg\" (UID: \"26f17329-55b8-4de4-82ee-98775dafa32a\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.627581 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26f17329-55b8-4de4-82ee-98775dafa32a-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg\" (UID: \"26f17329-55b8-4de4-82ee-98775dafa32a\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.658990 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzpfs\" (UniqueName: \"kubernetes.io/projected/26f17329-55b8-4de4-82ee-98775dafa32a-kube-api-access-pzpfs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg\" (UID: \"26f17329-55b8-4de4-82ee-98775dafa32a\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:21 crc kubenswrapper[4744]: I1008 09:39:21.726048 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:22 crc kubenswrapper[4744]: I1008 09:39:22.320187 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg"] Oct 08 09:39:23 crc kubenswrapper[4744]: I1008 09:39:23.327570 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" event={"ID":"26f17329-55b8-4de4-82ee-98775dafa32a","Type":"ContainerStarted","Data":"89a1061ef4c1b296eaa022b9c64167e6025c997579d6349ca258a18cac3d0b7f"} Oct 08 09:39:23 crc kubenswrapper[4744]: I1008 09:39:23.327903 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" event={"ID":"26f17329-55b8-4de4-82ee-98775dafa32a","Type":"ContainerStarted","Data":"6373a97c50ca7257719f38853236db85f4e46047f7b721271caca2108e0c6414"} Oct 08 09:39:23 crc kubenswrapper[4744]: I1008 09:39:23.356764 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" podStartSLOduration=1.96506912 podStartE2EDuration="2.356742396s" podCreationTimestamp="2025-10-08 09:39:21 +0000 UTC" firstStartedPulling="2025-10-08 09:39:22.330542027 +0000 UTC m=+1657.578187266" lastFinishedPulling="2025-10-08 09:39:22.722215303 +0000 UTC m=+1657.969860542" observedRunningTime="2025-10-08 09:39:23.353636999 +0000 UTC m=+1658.601282248" watchObservedRunningTime="2025-10-08 09:39:23.356742396 +0000 UTC m=+1658.604387635" Oct 08 09:39:27 crc kubenswrapper[4744]: I1008 09:39:27.364284 4744 generic.go:334] "Generic (PLEG): container finished" podID="26f17329-55b8-4de4-82ee-98775dafa32a" containerID="89a1061ef4c1b296eaa022b9c64167e6025c997579d6349ca258a18cac3d0b7f" exitCode=0 Oct 08 09:39:27 crc kubenswrapper[4744]: I1008 09:39:27.364393 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" event={"ID":"26f17329-55b8-4de4-82ee-98775dafa32a","Type":"ContainerDied","Data":"89a1061ef4c1b296eaa022b9c64167e6025c997579d6349ca258a18cac3d0b7f"} Oct 08 09:39:28 crc kubenswrapper[4744]: I1008 09:39:28.821724 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:28 crc kubenswrapper[4744]: I1008 09:39:28.914563 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzpfs\" (UniqueName: \"kubernetes.io/projected/26f17329-55b8-4de4-82ee-98775dafa32a-kube-api-access-pzpfs\") pod \"26f17329-55b8-4de4-82ee-98775dafa32a\" (UID: \"26f17329-55b8-4de4-82ee-98775dafa32a\") " Oct 08 09:39:28 crc kubenswrapper[4744]: I1008 09:39:28.914626 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26f17329-55b8-4de4-82ee-98775dafa32a-ssh-key\") pod \"26f17329-55b8-4de4-82ee-98775dafa32a\" (UID: \"26f17329-55b8-4de4-82ee-98775dafa32a\") " Oct 08 09:39:28 crc kubenswrapper[4744]: I1008 09:39:28.914712 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26f17329-55b8-4de4-82ee-98775dafa32a-inventory\") pod \"26f17329-55b8-4de4-82ee-98775dafa32a\" (UID: \"26f17329-55b8-4de4-82ee-98775dafa32a\") " Oct 08 09:39:28 crc kubenswrapper[4744]: I1008 09:39:28.928606 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f17329-55b8-4de4-82ee-98775dafa32a-kube-api-access-pzpfs" (OuterVolumeSpecName: "kube-api-access-pzpfs") pod "26f17329-55b8-4de4-82ee-98775dafa32a" (UID: "26f17329-55b8-4de4-82ee-98775dafa32a"). InnerVolumeSpecName "kube-api-access-pzpfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:39:28 crc kubenswrapper[4744]: I1008 09:39:28.944492 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26f17329-55b8-4de4-82ee-98775dafa32a-inventory" (OuterVolumeSpecName: "inventory") pod "26f17329-55b8-4de4-82ee-98775dafa32a" (UID: "26f17329-55b8-4de4-82ee-98775dafa32a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:39:28 crc kubenswrapper[4744]: I1008 09:39:28.954067 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26f17329-55b8-4de4-82ee-98775dafa32a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "26f17329-55b8-4de4-82ee-98775dafa32a" (UID: "26f17329-55b8-4de4-82ee-98775dafa32a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.016662 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzpfs\" (UniqueName: \"kubernetes.io/projected/26f17329-55b8-4de4-82ee-98775dafa32a-kube-api-access-pzpfs\") on node \"crc\" DevicePath \"\"" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.016699 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26f17329-55b8-4de4-82ee-98775dafa32a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.016708 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26f17329-55b8-4de4-82ee-98775dafa32a-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.390495 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" event={"ID":"26f17329-55b8-4de4-82ee-98775dafa32a","Type":"ContainerDied","Data":"6373a97c50ca7257719f38853236db85f4e46047f7b721271caca2108e0c6414"} Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.391087 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6373a97c50ca7257719f38853236db85f4e46047f7b721271caca2108e0c6414" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.390608 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.503705 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn"] Oct 08 09:39:29 crc kubenswrapper[4744]: E1008 09:39:29.504837 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f17329-55b8-4de4-82ee-98775dafa32a" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.504972 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f17329-55b8-4de4-82ee-98775dafa32a" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.509209 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f17329-55b8-4de4-82ee-98775dafa32a" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.510595 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.514505 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.514572 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.514757 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.514737 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.521207 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn"] Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.627045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c75eb05-f1da-4db0-87b6-7432e97f244c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llqtn\" (UID: \"0c75eb05-f1da-4db0-87b6-7432e97f244c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.627506 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c75eb05-f1da-4db0-87b6-7432e97f244c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llqtn\" (UID: \"0c75eb05-f1da-4db0-87b6-7432e97f244c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.627632 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdw8v\" (UniqueName: \"kubernetes.io/projected/0c75eb05-f1da-4db0-87b6-7432e97f244c-kube-api-access-fdw8v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llqtn\" (UID: \"0c75eb05-f1da-4db0-87b6-7432e97f244c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.729922 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c75eb05-f1da-4db0-87b6-7432e97f244c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llqtn\" (UID: \"0c75eb05-f1da-4db0-87b6-7432e97f244c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.730040 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdw8v\" (UniqueName: \"kubernetes.io/projected/0c75eb05-f1da-4db0-87b6-7432e97f244c-kube-api-access-fdw8v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llqtn\" (UID: \"0c75eb05-f1da-4db0-87b6-7432e97f244c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.730107 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c75eb05-f1da-4db0-87b6-7432e97f244c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llqtn\" (UID: \"0c75eb05-f1da-4db0-87b6-7432e97f244c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.735132 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c75eb05-f1da-4db0-87b6-7432e97f244c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llqtn\" (UID: \"0c75eb05-f1da-4db0-87b6-7432e97f244c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.735690 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c75eb05-f1da-4db0-87b6-7432e97f244c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llqtn\" (UID: \"0c75eb05-f1da-4db0-87b6-7432e97f244c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.752972 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdw8v\" (UniqueName: \"kubernetes.io/projected/0c75eb05-f1da-4db0-87b6-7432e97f244c-kube-api-access-fdw8v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-llqtn\" (UID: \"0c75eb05-f1da-4db0-87b6-7432e97f244c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:39:29 crc kubenswrapper[4744]: I1008 09:39:29.840188 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:39:30 crc kubenswrapper[4744]: I1008 09:39:30.420486 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn"] Oct 08 09:39:30 crc kubenswrapper[4744]: I1008 09:39:30.452997 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:39:30 crc kubenswrapper[4744]: E1008 09:39:30.453300 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:39:31 crc kubenswrapper[4744]: I1008 09:39:31.413115 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" event={"ID":"0c75eb05-f1da-4db0-87b6-7432e97f244c","Type":"ContainerStarted","Data":"97478bbac68c97a342195072be4524df21c40d0ff001858bfc07b16e281d4b9a"} Oct 08 09:39:31 crc kubenswrapper[4744]: I1008 09:39:31.414194 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" event={"ID":"0c75eb05-f1da-4db0-87b6-7432e97f244c","Type":"ContainerStarted","Data":"f39a7b5a75fad2a908a4bbf16b565bd34117ccef184ac0350894099878813529"} Oct 08 09:39:31 crc kubenswrapper[4744]: I1008 09:39:31.432232 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" podStartSLOduration=1.878898354 podStartE2EDuration="2.432199498s" podCreationTimestamp="2025-10-08 09:39:29 +0000 UTC" firstStartedPulling="2025-10-08 09:39:30.432552954 +0000 UTC m=+1665.680198203" lastFinishedPulling="2025-10-08 09:39:30.985854088 +0000 UTC m=+1666.233499347" observedRunningTime="2025-10-08 09:39:31.429504863 +0000 UTC m=+1666.677150102" watchObservedRunningTime="2025-10-08 09:39:31.432199498 +0000 UTC m=+1666.679844757" Oct 08 09:39:37 crc kubenswrapper[4744]: I1008 09:39:37.055408 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-q99xn"] Oct 08 09:39:37 crc kubenswrapper[4744]: I1008 09:39:37.063641 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-q99xn"] Oct 08 09:39:37 crc kubenswrapper[4744]: I1008 09:39:37.466633 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0779920-88cf-4475-9c77-5de26f893b6b" path="/var/lib/kubelet/pods/f0779920-88cf-4475-9c77-5de26f893b6b/volumes" Oct 08 09:39:43 crc kubenswrapper[4744]: I1008 09:39:43.032550 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-cmhgp"] Oct 08 09:39:43 crc kubenswrapper[4744]: I1008 09:39:43.042485 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-58rvm"] Oct 08 09:39:43 crc kubenswrapper[4744]: I1008 09:39:43.051290 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-cmhgp"] Oct 08 09:39:43 crc kubenswrapper[4744]: I1008 09:39:43.088960 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-58rvm"] Oct 08 09:39:43 crc kubenswrapper[4744]: I1008 09:39:43.465061 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56496cc2-c5e8-426d-8f7c-fbd84f7b266c" path="/var/lib/kubelet/pods/56496cc2-c5e8-426d-8f7c-fbd84f7b266c/volumes" Oct 08 09:39:43 crc kubenswrapper[4744]: I1008 09:39:43.465860 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daa62f04-afb7-4d1e-9218-cbc2cc22e807" path="/var/lib/kubelet/pods/daa62f04-afb7-4d1e-9218-cbc2cc22e807/volumes" Oct 08 09:39:44 crc kubenswrapper[4744]: I1008 09:39:44.038834 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-plnkl"] Oct 08 09:39:44 crc kubenswrapper[4744]: I1008 09:39:44.051013 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-plnkl"] Oct 08 09:39:44 crc kubenswrapper[4744]: I1008 09:39:44.453167 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:39:44 crc kubenswrapper[4744]: E1008 09:39:44.453402 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:39:45 crc kubenswrapper[4744]: I1008 09:39:45.465879 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="929c71a8-07ed-475b-b150-e5a1d92e22c5" path="/var/lib/kubelet/pods/929c71a8-07ed-475b-b150-e5a1d92e22c5/volumes" Oct 08 09:39:53 crc kubenswrapper[4744]: I1008 09:39:53.036221 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-bc72-account-create-gfwls"] Oct 08 09:39:53 crc kubenswrapper[4744]: I1008 09:39:53.043238 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-bc72-account-create-gfwls"] Oct 08 09:39:53 crc kubenswrapper[4744]: I1008 09:39:53.464554 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2edcd16b-76ad-4a92-b75b-033b9edd0024" path="/var/lib/kubelet/pods/2edcd16b-76ad-4a92-b75b-033b9edd0024/volumes" Oct 08 09:39:54 crc kubenswrapper[4744]: I1008 09:39:54.033762 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-9b88-account-create-v6x5w"] Oct 08 09:39:54 crc kubenswrapper[4744]: I1008 09:39:54.042806 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-9b88-account-create-v6x5w"] Oct 08 09:39:54 crc kubenswrapper[4744]: I1008 09:39:54.325637 4744 scope.go:117] "RemoveContainer" containerID="452f48ca078c410657ab748a065813649452a099e3162cc9e97b35a02c5b1efc" Oct 08 09:39:54 crc kubenswrapper[4744]: I1008 09:39:54.360921 4744 scope.go:117] "RemoveContainer" containerID="3361f212ebd338ebbe7433cc11d50c31512677ad8628093354c6f994e160c388" Oct 08 09:39:54 crc kubenswrapper[4744]: I1008 09:39:54.419436 4744 scope.go:117] "RemoveContainer" containerID="59ea7fea62b38cc171ee35f64831b057b77bd883211935c7db096d4df481a75d" Oct 08 09:39:54 crc kubenswrapper[4744]: I1008 09:39:54.480055 4744 scope.go:117] "RemoveContainer" containerID="a817ffeda72b2f03ed83fd5396bca4ad58202d5e6b39111d24b5e40c515ea1ad" Oct 08 09:39:54 crc kubenswrapper[4744]: I1008 09:39:54.516169 4744 scope.go:117] "RemoveContainer" containerID="bcd8e6d09ceff1ef8d99ee2137a4ec92fad6e768959c489f4623be6ecd49c7ce" Oct 08 09:39:54 crc kubenswrapper[4744]: I1008 09:39:54.616588 4744 scope.go:117] "RemoveContainer" containerID="01ab5fc768cb62b708dd9ca19f8ff741686c711b9bdd4095e355b80c2000bde9" Oct 08 09:39:55 crc kubenswrapper[4744]: I1008 09:39:55.471589 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7edf84ea-4d5a-4764-b6a7-ea03f6e312e4" path="/var/lib/kubelet/pods/7edf84ea-4d5a-4764-b6a7-ea03f6e312e4/volumes" Oct 08 09:39:56 crc kubenswrapper[4744]: I1008 09:39:56.455341 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:39:56 crc kubenswrapper[4744]: E1008 09:39:56.455939 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:40:03 crc kubenswrapper[4744]: I1008 09:40:03.081625 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-66f4-account-create-5dh4v"] Oct 08 09:40:03 crc kubenswrapper[4744]: I1008 09:40:03.130871 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-66f4-account-create-5dh4v"] Oct 08 09:40:03 crc kubenswrapper[4744]: I1008 09:40:03.463358 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="800eea78-e2f9-40d8-a2f9-bfe60b9adea9" path="/var/lib/kubelet/pods/800eea78-e2f9-40d8-a2f9-bfe60b9adea9/volumes" Oct 08 09:40:11 crc kubenswrapper[4744]: I1008 09:40:11.453780 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:40:11 crc kubenswrapper[4744]: E1008 09:40:11.454784 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:40:24 crc kubenswrapper[4744]: I1008 09:40:24.453395 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:40:24 crc kubenswrapper[4744]: E1008 09:40:24.454128 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:40:25 crc kubenswrapper[4744]: I1008 09:40:25.029004 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4txff"] Oct 08 09:40:25 crc kubenswrapper[4744]: I1008 09:40:25.042481 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-4txff"] Oct 08 09:40:25 crc kubenswrapper[4744]: I1008 09:40:25.464007 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93968d7d-7816-4580-8a04-f4ccacd584b6" path="/var/lib/kubelet/pods/93968d7d-7816-4580-8a04-f4ccacd584b6/volumes" Oct 08 09:40:26 crc kubenswrapper[4744]: I1008 09:40:26.956037 4744 generic.go:334] "Generic (PLEG): container finished" podID="0c75eb05-f1da-4db0-87b6-7432e97f244c" containerID="97478bbac68c97a342195072be4524df21c40d0ff001858bfc07b16e281d4b9a" exitCode=2 Oct 08 09:40:26 crc kubenswrapper[4744]: I1008 09:40:26.957149 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" event={"ID":"0c75eb05-f1da-4db0-87b6-7432e97f244c","Type":"ContainerDied","Data":"97478bbac68c97a342195072be4524df21c40d0ff001858bfc07b16e281d4b9a"} Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.385320 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.454986 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdw8v\" (UniqueName: \"kubernetes.io/projected/0c75eb05-f1da-4db0-87b6-7432e97f244c-kube-api-access-fdw8v\") pod \"0c75eb05-f1da-4db0-87b6-7432e97f244c\" (UID: \"0c75eb05-f1da-4db0-87b6-7432e97f244c\") " Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.455293 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c75eb05-f1da-4db0-87b6-7432e97f244c-ssh-key\") pod \"0c75eb05-f1da-4db0-87b6-7432e97f244c\" (UID: \"0c75eb05-f1da-4db0-87b6-7432e97f244c\") " Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.455441 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c75eb05-f1da-4db0-87b6-7432e97f244c-inventory\") pod \"0c75eb05-f1da-4db0-87b6-7432e97f244c\" (UID: \"0c75eb05-f1da-4db0-87b6-7432e97f244c\") " Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.462841 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c75eb05-f1da-4db0-87b6-7432e97f244c-kube-api-access-fdw8v" (OuterVolumeSpecName: "kube-api-access-fdw8v") pod "0c75eb05-f1da-4db0-87b6-7432e97f244c" (UID: "0c75eb05-f1da-4db0-87b6-7432e97f244c"). InnerVolumeSpecName "kube-api-access-fdw8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.494314 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c75eb05-f1da-4db0-87b6-7432e97f244c-inventory" (OuterVolumeSpecName: "inventory") pod "0c75eb05-f1da-4db0-87b6-7432e97f244c" (UID: "0c75eb05-f1da-4db0-87b6-7432e97f244c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.496741 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c75eb05-f1da-4db0-87b6-7432e97f244c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0c75eb05-f1da-4db0-87b6-7432e97f244c" (UID: "0c75eb05-f1da-4db0-87b6-7432e97f244c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.558073 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c75eb05-f1da-4db0-87b6-7432e97f244c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.558132 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c75eb05-f1da-4db0-87b6-7432e97f244c-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.558143 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdw8v\" (UniqueName: \"kubernetes.io/projected/0c75eb05-f1da-4db0-87b6-7432e97f244c-kube-api-access-fdw8v\") on node \"crc\" DevicePath \"\"" Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.979118 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" event={"ID":"0c75eb05-f1da-4db0-87b6-7432e97f244c","Type":"ContainerDied","Data":"f39a7b5a75fad2a908a4bbf16b565bd34117ccef184ac0350894099878813529"} Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.979190 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f39a7b5a75fad2a908a4bbf16b565bd34117ccef184ac0350894099878813529" Oct 08 09:40:28 crc kubenswrapper[4744]: I1008 09:40:28.979202 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.038758 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p"] Oct 08 09:40:36 crc kubenswrapper[4744]: E1008 09:40:36.039507 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c75eb05-f1da-4db0-87b6-7432e97f244c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.039519 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c75eb05-f1da-4db0-87b6-7432e97f244c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.039707 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c75eb05-f1da-4db0-87b6-7432e97f244c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.040328 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.043639 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.043639 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.043684 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.043710 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.051573 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p"] Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.137922 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6kwt\" (UniqueName: \"kubernetes.io/projected/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-kube-api-access-z6kwt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-48p2p\" (UID: \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.137974 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-48p2p\" (UID: \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.138407 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-48p2p\" (UID: \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.246353 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-48p2p\" (UID: \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.246467 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6kwt\" (UniqueName: \"kubernetes.io/projected/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-kube-api-access-z6kwt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-48p2p\" (UID: \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.246495 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-48p2p\" (UID: \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.267035 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-48p2p\" (UID: \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.277283 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6kwt\" (UniqueName: \"kubernetes.io/projected/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-kube-api-access-z6kwt\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-48p2p\" (UID: \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.277439 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-48p2p\" (UID: \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.363145 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:40:36 crc kubenswrapper[4744]: I1008 09:40:36.918865 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p"] Oct 08 09:40:37 crc kubenswrapper[4744]: I1008 09:40:37.103346 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" event={"ID":"6402ae4a-8d4f-4011-9a28-16b8b8303ba2","Type":"ContainerStarted","Data":"1123c3cfc1cb195b1d516c53d9740cd85e60ffad22438332e71abc655c5868f2"} Oct 08 09:40:38 crc kubenswrapper[4744]: I1008 09:40:38.116831 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" event={"ID":"6402ae4a-8d4f-4011-9a28-16b8b8303ba2","Type":"ContainerStarted","Data":"2c1bc45b415c831d02f110d8022cf5178bc6d2438fb903f0b5eef8497be69431"} Oct 08 09:40:38 crc kubenswrapper[4744]: I1008 09:40:38.136750 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" podStartSLOduration=1.62610631 podStartE2EDuration="2.136731413s" podCreationTimestamp="2025-10-08 09:40:36 +0000 UTC" firstStartedPulling="2025-10-08 09:40:36.936795934 +0000 UTC m=+1732.184441213" lastFinishedPulling="2025-10-08 09:40:37.447421077 +0000 UTC m=+1732.695066316" observedRunningTime="2025-10-08 09:40:38.131880127 +0000 UTC m=+1733.379525366" watchObservedRunningTime="2025-10-08 09:40:38.136731413 +0000 UTC m=+1733.384376652" Oct 08 09:40:38 crc kubenswrapper[4744]: I1008 09:40:38.453829 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:40:38 crc kubenswrapper[4744]: E1008 09:40:38.454149 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:40:50 crc kubenswrapper[4744]: I1008 09:40:50.047776 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-x4khv"] Oct 08 09:40:50 crc kubenswrapper[4744]: I1008 09:40:50.057271 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-x4khv"] Oct 08 09:40:51 crc kubenswrapper[4744]: I1008 09:40:51.466710 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8711d76-061a-4b5a-a2f6-05996808b79d" path="/var/lib/kubelet/pods/f8711d76-061a-4b5a-a2f6-05996808b79d/volumes" Oct 08 09:40:52 crc kubenswrapper[4744]: I1008 09:40:52.029396 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vgxfh"] Oct 08 09:40:52 crc kubenswrapper[4744]: I1008 09:40:52.035980 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vgxfh"] Oct 08 09:40:52 crc kubenswrapper[4744]: I1008 09:40:52.453197 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:40:52 crc kubenswrapper[4744]: E1008 09:40:52.453993 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:40:53 crc kubenswrapper[4744]: I1008 09:40:53.465645 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bc195e6-d464-4837-b6a0-f11b629e3446" path="/var/lib/kubelet/pods/6bc195e6-d464-4837-b6a0-f11b629e3446/volumes" Oct 08 09:40:54 crc kubenswrapper[4744]: I1008 09:40:54.749579 4744 scope.go:117] "RemoveContainer" containerID="b7479abcfb5a3ab44ee36628c757a9a869e08ea40da65001da7a1f10cd8b48ab" Oct 08 09:40:54 crc kubenswrapper[4744]: I1008 09:40:54.847545 4744 scope.go:117] "RemoveContainer" containerID="5d4430a6d79b848025ab9dc8f42b5fafa7e4cf074ebf7e7b5ce0acec938fecd9" Oct 08 09:40:54 crc kubenswrapper[4744]: I1008 09:40:54.870648 4744 scope.go:117] "RemoveContainer" containerID="398904171cafba1a279cf46661a2150313ab50fe0b08ce7ff92fbab6592beb76" Oct 08 09:40:54 crc kubenswrapper[4744]: I1008 09:40:54.939277 4744 scope.go:117] "RemoveContainer" containerID="2f8b94fa09ec47e8aeb6aab803b6abe46eea297c395a4af7d144259c22f62236" Oct 08 09:40:54 crc kubenswrapper[4744]: I1008 09:40:54.988793 4744 scope.go:117] "RemoveContainer" containerID="5aa01bfa090d80a1d857280f83be8e938f84d2155b94a373a4aa6fea1e452704" Oct 08 09:41:07 crc kubenswrapper[4744]: I1008 09:41:07.453122 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:41:07 crc kubenswrapper[4744]: E1008 09:41:07.453955 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:41:22 crc kubenswrapper[4744]: I1008 09:41:22.453791 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:41:22 crc kubenswrapper[4744]: E1008 09:41:22.455800 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:41:25 crc kubenswrapper[4744]: I1008 09:41:25.593052 4744 generic.go:334] "Generic (PLEG): container finished" podID="6402ae4a-8d4f-4011-9a28-16b8b8303ba2" containerID="2c1bc45b415c831d02f110d8022cf5178bc6d2438fb903f0b5eef8497be69431" exitCode=0 Oct 08 09:41:25 crc kubenswrapper[4744]: I1008 09:41:25.593258 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" event={"ID":"6402ae4a-8d4f-4011-9a28-16b8b8303ba2","Type":"ContainerDied","Data":"2c1bc45b415c831d02f110d8022cf5178bc6d2438fb903f0b5eef8497be69431"} Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.027172 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.071584 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-inventory\") pod \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\" (UID: \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\") " Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.071701 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6kwt\" (UniqueName: \"kubernetes.io/projected/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-kube-api-access-z6kwt\") pod \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\" (UID: \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\") " Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.071879 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-ssh-key\") pod \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\" (UID: \"6402ae4a-8d4f-4011-9a28-16b8b8303ba2\") " Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.080877 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-kube-api-access-z6kwt" (OuterVolumeSpecName: "kube-api-access-z6kwt") pod "6402ae4a-8d4f-4011-9a28-16b8b8303ba2" (UID: "6402ae4a-8d4f-4011-9a28-16b8b8303ba2"). InnerVolumeSpecName "kube-api-access-z6kwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.116709 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6402ae4a-8d4f-4011-9a28-16b8b8303ba2" (UID: "6402ae4a-8d4f-4011-9a28-16b8b8303ba2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.120327 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-inventory" (OuterVolumeSpecName: "inventory") pod "6402ae4a-8d4f-4011-9a28-16b8b8303ba2" (UID: "6402ae4a-8d4f-4011-9a28-16b8b8303ba2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.176170 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.176220 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6kwt\" (UniqueName: \"kubernetes.io/projected/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-kube-api-access-z6kwt\") on node \"crc\" DevicePath \"\"" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.176236 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6402ae4a-8d4f-4011-9a28-16b8b8303ba2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.613909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" event={"ID":"6402ae4a-8d4f-4011-9a28-16b8b8303ba2","Type":"ContainerDied","Data":"1123c3cfc1cb195b1d516c53d9740cd85e60ffad22438332e71abc655c5868f2"} Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.614201 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1123c3cfc1cb195b1d516c53d9740cd85e60ffad22438332e71abc655c5868f2" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.614002 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.704881 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4jwsc"] Oct 08 09:41:27 crc kubenswrapper[4744]: E1008 09:41:27.705707 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6402ae4a-8d4f-4011-9a28-16b8b8303ba2" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.705787 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6402ae4a-8d4f-4011-9a28-16b8b8303ba2" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.706064 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6402ae4a-8d4f-4011-9a28-16b8b8303ba2" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.707635 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.715193 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.724941 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4jwsc"] Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.726143 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.726179 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.726465 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.789640 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/606f778d-ee4d-4e8a-8cfd-d2dc86874016-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4jwsc\" (UID: \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\") " pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.789833 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htch4\" (UniqueName: \"kubernetes.io/projected/606f778d-ee4d-4e8a-8cfd-d2dc86874016-kube-api-access-htch4\") pod \"ssh-known-hosts-edpm-deployment-4jwsc\" (UID: \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\") " pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.789878 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/606f778d-ee4d-4e8a-8cfd-d2dc86874016-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4jwsc\" (UID: \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\") " pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.890902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htch4\" (UniqueName: \"kubernetes.io/projected/606f778d-ee4d-4e8a-8cfd-d2dc86874016-kube-api-access-htch4\") pod \"ssh-known-hosts-edpm-deployment-4jwsc\" (UID: \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\") " pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.890995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/606f778d-ee4d-4e8a-8cfd-d2dc86874016-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4jwsc\" (UID: \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\") " pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.891046 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/606f778d-ee4d-4e8a-8cfd-d2dc86874016-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4jwsc\" (UID: \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\") " pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.897129 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/606f778d-ee4d-4e8a-8cfd-d2dc86874016-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-4jwsc\" (UID: \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\") " pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.897662 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/606f778d-ee4d-4e8a-8cfd-d2dc86874016-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-4jwsc\" (UID: \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\") " pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:27 crc kubenswrapper[4744]: I1008 09:41:27.911345 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htch4\" (UniqueName: \"kubernetes.io/projected/606f778d-ee4d-4e8a-8cfd-d2dc86874016-kube-api-access-htch4\") pod \"ssh-known-hosts-edpm-deployment-4jwsc\" (UID: \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\") " pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:28 crc kubenswrapper[4744]: I1008 09:41:28.035383 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:28 crc kubenswrapper[4744]: I1008 09:41:28.586196 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4jwsc"] Oct 08 09:41:28 crc kubenswrapper[4744]: I1008 09:41:28.625931 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" event={"ID":"606f778d-ee4d-4e8a-8cfd-d2dc86874016","Type":"ContainerStarted","Data":"1738a464497550ae674c66a81e7a5c49382a2a32577efefe9458273ece343d2c"} Oct 08 09:41:29 crc kubenswrapper[4744]: I1008 09:41:29.636864 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" event={"ID":"606f778d-ee4d-4e8a-8cfd-d2dc86874016","Type":"ContainerStarted","Data":"e87c8e3bb3965b78b6ec142165e56ed4fe5c61c052fd8f46119a17d9b77346d9"} Oct 08 09:41:34 crc kubenswrapper[4744]: I1008 09:41:34.046130 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" podStartSLOduration=6.59754818 podStartE2EDuration="7.046103881s" podCreationTimestamp="2025-10-08 09:41:27 +0000 UTC" firstStartedPulling="2025-10-08 09:41:28.606590886 +0000 UTC m=+1783.854236125" lastFinishedPulling="2025-10-08 09:41:29.055146587 +0000 UTC m=+1784.302791826" observedRunningTime="2025-10-08 09:41:29.656979404 +0000 UTC m=+1784.904624763" watchObservedRunningTime="2025-10-08 09:41:34.046103881 +0000 UTC m=+1789.293749120" Oct 08 09:41:34 crc kubenswrapper[4744]: I1008 09:41:34.051502 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-62khr"] Oct 08 09:41:34 crc kubenswrapper[4744]: I1008 09:41:34.070085 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-62khr"] Oct 08 09:41:34 crc kubenswrapper[4744]: I1008 09:41:34.454056 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:41:34 crc kubenswrapper[4744]: E1008 09:41:34.454501 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:41:35 crc kubenswrapper[4744]: I1008 09:41:35.466199 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc006ede-60d5-417e-bde3-e3bca31b912b" path="/var/lib/kubelet/pods/fc006ede-60d5-417e-bde3-e3bca31b912b/volumes" Oct 08 09:41:36 crc kubenswrapper[4744]: I1008 09:41:36.704299 4744 generic.go:334] "Generic (PLEG): container finished" podID="606f778d-ee4d-4e8a-8cfd-d2dc86874016" containerID="e87c8e3bb3965b78b6ec142165e56ed4fe5c61c052fd8f46119a17d9b77346d9" exitCode=0 Oct 08 09:41:36 crc kubenswrapper[4744]: I1008 09:41:36.704402 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" event={"ID":"606f778d-ee4d-4e8a-8cfd-d2dc86874016","Type":"ContainerDied","Data":"e87c8e3bb3965b78b6ec142165e56ed4fe5c61c052fd8f46119a17d9b77346d9"} Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.089810 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.141203 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/606f778d-ee4d-4e8a-8cfd-d2dc86874016-inventory-0\") pod \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\" (UID: \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\") " Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.141324 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/606f778d-ee4d-4e8a-8cfd-d2dc86874016-ssh-key-openstack-edpm-ipam\") pod \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\" (UID: \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\") " Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.141446 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htch4\" (UniqueName: \"kubernetes.io/projected/606f778d-ee4d-4e8a-8cfd-d2dc86874016-kube-api-access-htch4\") pod \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\" (UID: \"606f778d-ee4d-4e8a-8cfd-d2dc86874016\") " Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.151207 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/606f778d-ee4d-4e8a-8cfd-d2dc86874016-kube-api-access-htch4" (OuterVolumeSpecName: "kube-api-access-htch4") pod "606f778d-ee4d-4e8a-8cfd-d2dc86874016" (UID: "606f778d-ee4d-4e8a-8cfd-d2dc86874016"). InnerVolumeSpecName "kube-api-access-htch4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.173191 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/606f778d-ee4d-4e8a-8cfd-d2dc86874016-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "606f778d-ee4d-4e8a-8cfd-d2dc86874016" (UID: "606f778d-ee4d-4e8a-8cfd-d2dc86874016"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.173809 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/606f778d-ee4d-4e8a-8cfd-d2dc86874016-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "606f778d-ee4d-4e8a-8cfd-d2dc86874016" (UID: "606f778d-ee4d-4e8a-8cfd-d2dc86874016"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.244022 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htch4\" (UniqueName: \"kubernetes.io/projected/606f778d-ee4d-4e8a-8cfd-d2dc86874016-kube-api-access-htch4\") on node \"crc\" DevicePath \"\"" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.244062 4744 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/606f778d-ee4d-4e8a-8cfd-d2dc86874016-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.244076 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/606f778d-ee4d-4e8a-8cfd-d2dc86874016-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.740501 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" event={"ID":"606f778d-ee4d-4e8a-8cfd-d2dc86874016","Type":"ContainerDied","Data":"1738a464497550ae674c66a81e7a5c49382a2a32577efefe9458273ece343d2c"} Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.740552 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1738a464497550ae674c66a81e7a5c49382a2a32577efefe9458273ece343d2c" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.740619 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-4jwsc" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.829198 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc"] Oct 08 09:41:38 crc kubenswrapper[4744]: E1008 09:41:38.830299 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="606f778d-ee4d-4e8a-8cfd-d2dc86874016" containerName="ssh-known-hosts-edpm-deployment" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.830322 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="606f778d-ee4d-4e8a-8cfd-d2dc86874016" containerName="ssh-known-hosts-edpm-deployment" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.830585 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="606f778d-ee4d-4e8a-8cfd-d2dc86874016" containerName="ssh-known-hosts-edpm-deployment" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.831404 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.834598 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.835096 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.835440 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.847707 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.852553 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc"] Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.959153 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q28jc\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.959247 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q28jc\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:38 crc kubenswrapper[4744]: I1008 09:41:38.959298 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25bjt\" (UniqueName: \"kubernetes.io/projected/3a036815-b901-4cd0-8ed9-b30c6c63be44-kube-api-access-25bjt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q28jc\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:39 crc kubenswrapper[4744]: I1008 09:41:39.061708 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q28jc\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:39 crc kubenswrapper[4744]: I1008 09:41:39.061762 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q28jc\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:39 crc kubenswrapper[4744]: I1008 09:41:39.061806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25bjt\" (UniqueName: \"kubernetes.io/projected/3a036815-b901-4cd0-8ed9-b30c6c63be44-kube-api-access-25bjt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q28jc\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:39 crc kubenswrapper[4744]: I1008 09:41:39.066156 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q28jc\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:39 crc kubenswrapper[4744]: I1008 09:41:39.072183 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q28jc\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:39 crc kubenswrapper[4744]: I1008 09:41:39.082285 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25bjt\" (UniqueName: \"kubernetes.io/projected/3a036815-b901-4cd0-8ed9-b30c6c63be44-kube-api-access-25bjt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-q28jc\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:39 crc kubenswrapper[4744]: I1008 09:41:39.151811 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:39 crc kubenswrapper[4744]: I1008 09:41:39.751708 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc"] Oct 08 09:41:39 crc kubenswrapper[4744]: I1008 09:41:39.823564 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" event={"ID":"3a036815-b901-4cd0-8ed9-b30c6c63be44","Type":"ContainerStarted","Data":"2b3eac17dfbc56889462272bf2250272cbcb79620e7728383be0b8882ba807ac"} Oct 08 09:41:40 crc kubenswrapper[4744]: I1008 09:41:40.837554 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" event={"ID":"3a036815-b901-4cd0-8ed9-b30c6c63be44","Type":"ContainerStarted","Data":"a875f357e53d04f7ad56db08a071dc57998c61748d053575531c0fa580986d56"} Oct 08 09:41:40 crc kubenswrapper[4744]: I1008 09:41:40.867808 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" podStartSLOduration=2.341240756 podStartE2EDuration="2.867775742s" podCreationTimestamp="2025-10-08 09:41:38 +0000 UTC" firstStartedPulling="2025-10-08 09:41:39.723036382 +0000 UTC m=+1794.970681621" lastFinishedPulling="2025-10-08 09:41:40.249571338 +0000 UTC m=+1795.497216607" observedRunningTime="2025-10-08 09:41:40.85514439 +0000 UTC m=+1796.102789639" watchObservedRunningTime="2025-10-08 09:41:40.867775742 +0000 UTC m=+1796.115421021" Oct 08 09:41:48 crc kubenswrapper[4744]: I1008 09:41:48.453271 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:41:48 crc kubenswrapper[4744]: E1008 09:41:48.453960 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:41:48 crc kubenswrapper[4744]: I1008 09:41:48.905035 4744 generic.go:334] "Generic (PLEG): container finished" podID="3a036815-b901-4cd0-8ed9-b30c6c63be44" containerID="a875f357e53d04f7ad56db08a071dc57998c61748d053575531c0fa580986d56" exitCode=0 Oct 08 09:41:48 crc kubenswrapper[4744]: I1008 09:41:48.905106 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" event={"ID":"3a036815-b901-4cd0-8ed9-b30c6c63be44","Type":"ContainerDied","Data":"a875f357e53d04f7ad56db08a071dc57998c61748d053575531c0fa580986d56"} Oct 08 09:41:50 crc kubenswrapper[4744]: I1008 09:41:50.320335 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:50 crc kubenswrapper[4744]: I1008 09:41:50.415447 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-inventory\") pod \"3a036815-b901-4cd0-8ed9-b30c6c63be44\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " Oct 08 09:41:50 crc kubenswrapper[4744]: I1008 09:41:50.415704 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-ssh-key\") pod \"3a036815-b901-4cd0-8ed9-b30c6c63be44\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " Oct 08 09:41:50 crc kubenswrapper[4744]: I1008 09:41:50.415770 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25bjt\" (UniqueName: \"kubernetes.io/projected/3a036815-b901-4cd0-8ed9-b30c6c63be44-kube-api-access-25bjt\") pod \"3a036815-b901-4cd0-8ed9-b30c6c63be44\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " Oct 08 09:41:50 crc kubenswrapper[4744]: I1008 09:41:50.424837 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a036815-b901-4cd0-8ed9-b30c6c63be44-kube-api-access-25bjt" (OuterVolumeSpecName: "kube-api-access-25bjt") pod "3a036815-b901-4cd0-8ed9-b30c6c63be44" (UID: "3a036815-b901-4cd0-8ed9-b30c6c63be44"). InnerVolumeSpecName "kube-api-access-25bjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:41:50 crc kubenswrapper[4744]: E1008 09:41:50.449680 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-ssh-key podName:3a036815-b901-4cd0-8ed9-b30c6c63be44 nodeName:}" failed. No retries permitted until 2025-10-08 09:41:50.949644575 +0000 UTC m=+1806.197289824 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-ssh-key") pod "3a036815-b901-4cd0-8ed9-b30c6c63be44" (UID: "3a036815-b901-4cd0-8ed9-b30c6c63be44") : error deleting /var/lib/kubelet/pods/3a036815-b901-4cd0-8ed9-b30c6c63be44/volume-subpaths: remove /var/lib/kubelet/pods/3a036815-b901-4cd0-8ed9-b30c6c63be44/volume-subpaths: no such file or directory Oct 08 09:41:50 crc kubenswrapper[4744]: I1008 09:41:50.453437 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-inventory" (OuterVolumeSpecName: "inventory") pod "3a036815-b901-4cd0-8ed9-b30c6c63be44" (UID: "3a036815-b901-4cd0-8ed9-b30c6c63be44"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:41:50 crc kubenswrapper[4744]: I1008 09:41:50.518012 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25bjt\" (UniqueName: \"kubernetes.io/projected/3a036815-b901-4cd0-8ed9-b30c6c63be44-kube-api-access-25bjt\") on node \"crc\" DevicePath \"\"" Oct 08 09:41:50 crc kubenswrapper[4744]: I1008 09:41:50.518050 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:41:50 crc kubenswrapper[4744]: I1008 09:41:50.926335 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" event={"ID":"3a036815-b901-4cd0-8ed9-b30c6c63be44","Type":"ContainerDied","Data":"2b3eac17dfbc56889462272bf2250272cbcb79620e7728383be0b8882ba807ac"} Oct 08 09:41:50 crc kubenswrapper[4744]: I1008 09:41:50.926751 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b3eac17dfbc56889462272bf2250272cbcb79620e7728383be0b8882ba807ac" Oct 08 09:41:50 crc kubenswrapper[4744]: I1008 09:41:50.926450 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.018118 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7"] Oct 08 09:41:51 crc kubenswrapper[4744]: E1008 09:41:51.018721 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a036815-b901-4cd0-8ed9-b30c6c63be44" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.018748 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a036815-b901-4cd0-8ed9-b30c6c63be44" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.019351 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a036815-b901-4cd0-8ed9-b30c6c63be44" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.020342 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.028245 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-ssh-key\") pod \"3a036815-b901-4cd0-8ed9-b30c6c63be44\" (UID: \"3a036815-b901-4cd0-8ed9-b30c6c63be44\") " Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.032998 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3a036815-b901-4cd0-8ed9-b30c6c63be44" (UID: "3a036815-b901-4cd0-8ed9-b30c6c63be44"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.045763 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7"] Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.130444 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a08d7867-ad28-4079-a99d-f6db36d9a323-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7\" (UID: \"a08d7867-ad28-4079-a99d-f6db36d9a323\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.130521 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lbpp\" (UniqueName: \"kubernetes.io/projected/a08d7867-ad28-4079-a99d-f6db36d9a323-kube-api-access-2lbpp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7\" (UID: \"a08d7867-ad28-4079-a99d-f6db36d9a323\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.130546 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a08d7867-ad28-4079-a99d-f6db36d9a323-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7\" (UID: \"a08d7867-ad28-4079-a99d-f6db36d9a323\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.130648 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a036815-b901-4cd0-8ed9-b30c6c63be44-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.233249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a08d7867-ad28-4079-a99d-f6db36d9a323-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7\" (UID: \"a08d7867-ad28-4079-a99d-f6db36d9a323\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.233383 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lbpp\" (UniqueName: \"kubernetes.io/projected/a08d7867-ad28-4079-a99d-f6db36d9a323-kube-api-access-2lbpp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7\" (UID: \"a08d7867-ad28-4079-a99d-f6db36d9a323\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.233421 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a08d7867-ad28-4079-a99d-f6db36d9a323-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7\" (UID: \"a08d7867-ad28-4079-a99d-f6db36d9a323\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.238992 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a08d7867-ad28-4079-a99d-f6db36d9a323-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7\" (UID: \"a08d7867-ad28-4079-a99d-f6db36d9a323\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.245983 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a08d7867-ad28-4079-a99d-f6db36d9a323-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7\" (UID: \"a08d7867-ad28-4079-a99d-f6db36d9a323\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.253096 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lbpp\" (UniqueName: \"kubernetes.io/projected/a08d7867-ad28-4079-a99d-f6db36d9a323-kube-api-access-2lbpp\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7\" (UID: \"a08d7867-ad28-4079-a99d-f6db36d9a323\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.409535 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:41:51 crc kubenswrapper[4744]: I1008 09:41:51.991138 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7"] Oct 08 09:41:52 crc kubenswrapper[4744]: I1008 09:41:52.949587 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" event={"ID":"a08d7867-ad28-4079-a99d-f6db36d9a323","Type":"ContainerStarted","Data":"7c29bfe4edffea256e1fdc8047cb22421a81ce926b13da96a82073a4dcd53a25"} Oct 08 09:41:52 crc kubenswrapper[4744]: I1008 09:41:52.949645 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" event={"ID":"a08d7867-ad28-4079-a99d-f6db36d9a323","Type":"ContainerStarted","Data":"77846feaf385a2c7acef9497ce6ace42b2eb33eae7b01b0cfe66b8bba1120be1"} Oct 08 09:41:52 crc kubenswrapper[4744]: I1008 09:41:52.968492 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" podStartSLOduration=2.522403208 podStartE2EDuration="2.968472088s" podCreationTimestamp="2025-10-08 09:41:50 +0000 UTC" firstStartedPulling="2025-10-08 09:41:52.001069227 +0000 UTC m=+1807.248714466" lastFinishedPulling="2025-10-08 09:41:52.447138067 +0000 UTC m=+1807.694783346" observedRunningTime="2025-10-08 09:41:52.966423221 +0000 UTC m=+1808.214068460" watchObservedRunningTime="2025-10-08 09:41:52.968472088 +0000 UTC m=+1808.216117327" Oct 08 09:41:55 crc kubenswrapper[4744]: I1008 09:41:55.090291 4744 scope.go:117] "RemoveContainer" containerID="519f31b51a00c71469332f0ce2b05612c09ca86e54492eacd13c5767b8c460d8" Oct 08 09:42:03 crc kubenswrapper[4744]: I1008 09:42:03.064923 4744 generic.go:334] "Generic (PLEG): container finished" podID="a08d7867-ad28-4079-a99d-f6db36d9a323" containerID="7c29bfe4edffea256e1fdc8047cb22421a81ce926b13da96a82073a4dcd53a25" exitCode=0 Oct 08 09:42:03 crc kubenswrapper[4744]: I1008 09:42:03.065755 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" event={"ID":"a08d7867-ad28-4079-a99d-f6db36d9a323","Type":"ContainerDied","Data":"7c29bfe4edffea256e1fdc8047cb22421a81ce926b13da96a82073a4dcd53a25"} Oct 08 09:42:03 crc kubenswrapper[4744]: I1008 09:42:03.453757 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:42:03 crc kubenswrapper[4744]: E1008 09:42:03.454088 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:42:04 crc kubenswrapper[4744]: I1008 09:42:04.497569 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:42:04 crc kubenswrapper[4744]: I1008 09:42:04.663119 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a08d7867-ad28-4079-a99d-f6db36d9a323-inventory\") pod \"a08d7867-ad28-4079-a99d-f6db36d9a323\" (UID: \"a08d7867-ad28-4079-a99d-f6db36d9a323\") " Oct 08 09:42:04 crc kubenswrapper[4744]: I1008 09:42:04.663647 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a08d7867-ad28-4079-a99d-f6db36d9a323-ssh-key\") pod \"a08d7867-ad28-4079-a99d-f6db36d9a323\" (UID: \"a08d7867-ad28-4079-a99d-f6db36d9a323\") " Oct 08 09:42:04 crc kubenswrapper[4744]: I1008 09:42:04.663738 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lbpp\" (UniqueName: \"kubernetes.io/projected/a08d7867-ad28-4079-a99d-f6db36d9a323-kube-api-access-2lbpp\") pod \"a08d7867-ad28-4079-a99d-f6db36d9a323\" (UID: \"a08d7867-ad28-4079-a99d-f6db36d9a323\") " Oct 08 09:42:04 crc kubenswrapper[4744]: I1008 09:42:04.671813 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a08d7867-ad28-4079-a99d-f6db36d9a323-kube-api-access-2lbpp" (OuterVolumeSpecName: "kube-api-access-2lbpp") pod "a08d7867-ad28-4079-a99d-f6db36d9a323" (UID: "a08d7867-ad28-4079-a99d-f6db36d9a323"). InnerVolumeSpecName "kube-api-access-2lbpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:42:04 crc kubenswrapper[4744]: I1008 09:42:04.701018 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a08d7867-ad28-4079-a99d-f6db36d9a323-inventory" (OuterVolumeSpecName: "inventory") pod "a08d7867-ad28-4079-a99d-f6db36d9a323" (UID: "a08d7867-ad28-4079-a99d-f6db36d9a323"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:42:04 crc kubenswrapper[4744]: I1008 09:42:04.701959 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a08d7867-ad28-4079-a99d-f6db36d9a323-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a08d7867-ad28-4079-a99d-f6db36d9a323" (UID: "a08d7867-ad28-4079-a99d-f6db36d9a323"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:42:04 crc kubenswrapper[4744]: I1008 09:42:04.766046 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a08d7867-ad28-4079-a99d-f6db36d9a323-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:42:04 crc kubenswrapper[4744]: I1008 09:42:04.766094 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a08d7867-ad28-4079-a99d-f6db36d9a323-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:42:04 crc kubenswrapper[4744]: I1008 09:42:04.766108 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lbpp\" (UniqueName: \"kubernetes.io/projected/a08d7867-ad28-4079-a99d-f6db36d9a323-kube-api-access-2lbpp\") on node \"crc\" DevicePath \"\"" Oct 08 09:42:05 crc kubenswrapper[4744]: I1008 09:42:05.099915 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" event={"ID":"a08d7867-ad28-4079-a99d-f6db36d9a323","Type":"ContainerDied","Data":"77846feaf385a2c7acef9497ce6ace42b2eb33eae7b01b0cfe66b8bba1120be1"} Oct 08 09:42:05 crc kubenswrapper[4744]: I1008 09:42:05.099978 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77846feaf385a2c7acef9497ce6ace42b2eb33eae7b01b0cfe66b8bba1120be1" Oct 08 09:42:05 crc kubenswrapper[4744]: I1008 09:42:05.100080 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7" Oct 08 09:42:15 crc kubenswrapper[4744]: I1008 09:42:15.458164 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:42:15 crc kubenswrapper[4744]: E1008 09:42:15.460659 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:42:28 crc kubenswrapper[4744]: I1008 09:42:28.454498 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:42:28 crc kubenswrapper[4744]: E1008 09:42:28.455410 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:42:40 crc kubenswrapper[4744]: I1008 09:42:40.453511 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:42:40 crc kubenswrapper[4744]: E1008 09:42:40.454599 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:42:53 crc kubenswrapper[4744]: I1008 09:42:53.455322 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:42:54 crc kubenswrapper[4744]: I1008 09:42:54.655710 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"edfca7ab5f128667cb0425a5ce4b92f849206f6e3061358fe9d17c2c761c3416"} Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.562325 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bjf54"] Oct 08 09:44:17 crc kubenswrapper[4744]: E1008 09:44:17.563984 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a08d7867-ad28-4079-a99d-f6db36d9a323" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.564128 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a08d7867-ad28-4079-a99d-f6db36d9a323" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.564351 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a08d7867-ad28-4079-a99d-f6db36d9a323" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.565782 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.580227 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bjf54"] Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.754078 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-catalog-content\") pod \"redhat-operators-bjf54\" (UID: \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\") " pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.754157 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qg8r\" (UniqueName: \"kubernetes.io/projected/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-kube-api-access-2qg8r\") pod \"redhat-operators-bjf54\" (UID: \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\") " pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.754530 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-utilities\") pod \"redhat-operators-bjf54\" (UID: \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\") " pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.858036 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-utilities\") pod \"redhat-operators-bjf54\" (UID: \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\") " pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.858191 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-catalog-content\") pod \"redhat-operators-bjf54\" (UID: \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\") " pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.858230 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qg8r\" (UniqueName: \"kubernetes.io/projected/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-kube-api-access-2qg8r\") pod \"redhat-operators-bjf54\" (UID: \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\") " pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.859521 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-utilities\") pod \"redhat-operators-bjf54\" (UID: \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\") " pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.859860 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-catalog-content\") pod \"redhat-operators-bjf54\" (UID: \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\") " pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.887383 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qg8r\" (UniqueName: \"kubernetes.io/projected/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-kube-api-access-2qg8r\") pod \"redhat-operators-bjf54\" (UID: \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\") " pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:17 crc kubenswrapper[4744]: I1008 09:44:17.922990 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:18 crc kubenswrapper[4744]: I1008 09:44:18.478363 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bjf54"] Oct 08 09:44:19 crc kubenswrapper[4744]: I1008 09:44:19.420805 4744 generic.go:334] "Generic (PLEG): container finished" podID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" containerID="ff21ea129c72d75604fbd66bd3345829b7396c8ed8cc2f2719e0cb3e1f6d66c9" exitCode=0 Oct 08 09:44:19 crc kubenswrapper[4744]: I1008 09:44:19.420913 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjf54" event={"ID":"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86","Type":"ContainerDied","Data":"ff21ea129c72d75604fbd66bd3345829b7396c8ed8cc2f2719e0cb3e1f6d66c9"} Oct 08 09:44:19 crc kubenswrapper[4744]: I1008 09:44:19.421058 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjf54" event={"ID":"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86","Type":"ContainerStarted","Data":"71352f49fc80db96c3fa2f2b69690ad4e26d0428d56fddc9c279f469a6c1d9c3"} Oct 08 09:44:19 crc kubenswrapper[4744]: I1008 09:44:19.423837 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 09:44:20 crc kubenswrapper[4744]: I1008 09:44:20.434971 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjf54" event={"ID":"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86","Type":"ContainerStarted","Data":"6af2551c932d721213e066ac0357692c3d388633c65d9f53bdfd611b162c00c6"} Oct 08 09:44:24 crc kubenswrapper[4744]: I1008 09:44:24.465724 4744 generic.go:334] "Generic (PLEG): container finished" podID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" containerID="6af2551c932d721213e066ac0357692c3d388633c65d9f53bdfd611b162c00c6" exitCode=0 Oct 08 09:44:24 crc kubenswrapper[4744]: I1008 09:44:24.465815 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjf54" event={"ID":"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86","Type":"ContainerDied","Data":"6af2551c932d721213e066ac0357692c3d388633c65d9f53bdfd611b162c00c6"} Oct 08 09:44:25 crc kubenswrapper[4744]: I1008 09:44:25.477060 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjf54" event={"ID":"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86","Type":"ContainerStarted","Data":"172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d"} Oct 08 09:44:25 crc kubenswrapper[4744]: I1008 09:44:25.510780 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bjf54" podStartSLOduration=3.035939298 podStartE2EDuration="8.510755991s" podCreationTimestamp="2025-10-08 09:44:17 +0000 UTC" firstStartedPulling="2025-10-08 09:44:19.423538624 +0000 UTC m=+1954.671183863" lastFinishedPulling="2025-10-08 09:44:24.898355297 +0000 UTC m=+1960.146000556" observedRunningTime="2025-10-08 09:44:25.50353297 +0000 UTC m=+1960.751178239" watchObservedRunningTime="2025-10-08 09:44:25.510755991 +0000 UTC m=+1960.758401230" Oct 08 09:44:27 crc kubenswrapper[4744]: I1008 09:44:27.923613 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:27 crc kubenswrapper[4744]: I1008 09:44:27.925465 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:29 crc kubenswrapper[4744]: I1008 09:44:29.001988 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bjf54" podUID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" containerName="registry-server" probeResult="failure" output=< Oct 08 09:44:29 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 09:44:29 crc kubenswrapper[4744]: > Oct 08 09:44:37 crc kubenswrapper[4744]: I1008 09:44:37.969082 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:38 crc kubenswrapper[4744]: I1008 09:44:38.033633 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:38 crc kubenswrapper[4744]: I1008 09:44:38.213876 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bjf54"] Oct 08 09:44:39 crc kubenswrapper[4744]: I1008 09:44:39.609697 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bjf54" podUID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" containerName="registry-server" containerID="cri-o://172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d" gracePeriod=2 Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.110008 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.307744 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-catalog-content\") pod \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\" (UID: \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\") " Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.307960 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-utilities\") pod \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\" (UID: \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\") " Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.308047 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qg8r\" (UniqueName: \"kubernetes.io/projected/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-kube-api-access-2qg8r\") pod \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\" (UID: \"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86\") " Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.308742 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-utilities" (OuterVolumeSpecName: "utilities") pod "8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" (UID: "8c3a3397-3ed0-49fd-9ab4-4e6f81909a86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.313830 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-kube-api-access-2qg8r" (OuterVolumeSpecName: "kube-api-access-2qg8r") pod "8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" (UID: "8c3a3397-3ed0-49fd-9ab4-4e6f81909a86"). InnerVolumeSpecName "kube-api-access-2qg8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.397641 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" (UID: "8c3a3397-3ed0-49fd-9ab4-4e6f81909a86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.410017 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.410847 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qg8r\" (UniqueName: \"kubernetes.io/projected/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-kube-api-access-2qg8r\") on node \"crc\" DevicePath \"\"" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.410938 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.622620 4744 generic.go:334] "Generic (PLEG): container finished" podID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" containerID="172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d" exitCode=0 Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.622672 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjf54" event={"ID":"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86","Type":"ContainerDied","Data":"172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d"} Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.622712 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bjf54" event={"ID":"8c3a3397-3ed0-49fd-9ab4-4e6f81909a86","Type":"ContainerDied","Data":"71352f49fc80db96c3fa2f2b69690ad4e26d0428d56fddc9c279f469a6c1d9c3"} Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.622732 4744 scope.go:117] "RemoveContainer" containerID="172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.624014 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bjf54" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.659764 4744 scope.go:117] "RemoveContainer" containerID="6af2551c932d721213e066ac0357692c3d388633c65d9f53bdfd611b162c00c6" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.670353 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bjf54"] Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.677405 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bjf54"] Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.684812 4744 scope.go:117] "RemoveContainer" containerID="ff21ea129c72d75604fbd66bd3345829b7396c8ed8cc2f2719e0cb3e1f6d66c9" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.725553 4744 scope.go:117] "RemoveContainer" containerID="172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d" Oct 08 09:44:40 crc kubenswrapper[4744]: E1008 09:44:40.725989 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d\": container with ID starting with 172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d not found: ID does not exist" containerID="172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.726043 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d"} err="failed to get container status \"172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d\": rpc error: code = NotFound desc = could not find container \"172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d\": container with ID starting with 172a9861baa7b800689b5f09b9796c3fde5cd0f667ec6e1721f4d5c73e99814d not found: ID does not exist" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.726077 4744 scope.go:117] "RemoveContainer" containerID="6af2551c932d721213e066ac0357692c3d388633c65d9f53bdfd611b162c00c6" Oct 08 09:44:40 crc kubenswrapper[4744]: E1008 09:44:40.726582 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6af2551c932d721213e066ac0357692c3d388633c65d9f53bdfd611b162c00c6\": container with ID starting with 6af2551c932d721213e066ac0357692c3d388633c65d9f53bdfd611b162c00c6 not found: ID does not exist" containerID="6af2551c932d721213e066ac0357692c3d388633c65d9f53bdfd611b162c00c6" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.726614 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6af2551c932d721213e066ac0357692c3d388633c65d9f53bdfd611b162c00c6"} err="failed to get container status \"6af2551c932d721213e066ac0357692c3d388633c65d9f53bdfd611b162c00c6\": rpc error: code = NotFound desc = could not find container \"6af2551c932d721213e066ac0357692c3d388633c65d9f53bdfd611b162c00c6\": container with ID starting with 6af2551c932d721213e066ac0357692c3d388633c65d9f53bdfd611b162c00c6 not found: ID does not exist" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.726633 4744 scope.go:117] "RemoveContainer" containerID="ff21ea129c72d75604fbd66bd3345829b7396c8ed8cc2f2719e0cb3e1f6d66c9" Oct 08 09:44:40 crc kubenswrapper[4744]: E1008 09:44:40.726923 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff21ea129c72d75604fbd66bd3345829b7396c8ed8cc2f2719e0cb3e1f6d66c9\": container with ID starting with ff21ea129c72d75604fbd66bd3345829b7396c8ed8cc2f2719e0cb3e1f6d66c9 not found: ID does not exist" containerID="ff21ea129c72d75604fbd66bd3345829b7396c8ed8cc2f2719e0cb3e1f6d66c9" Oct 08 09:44:40 crc kubenswrapper[4744]: I1008 09:44:40.726957 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff21ea129c72d75604fbd66bd3345829b7396c8ed8cc2f2719e0cb3e1f6d66c9"} err="failed to get container status \"ff21ea129c72d75604fbd66bd3345829b7396c8ed8cc2f2719e0cb3e1f6d66c9\": rpc error: code = NotFound desc = could not find container \"ff21ea129c72d75604fbd66bd3345829b7396c8ed8cc2f2719e0cb3e1f6d66c9\": container with ID starting with ff21ea129c72d75604fbd66bd3345829b7396c8ed8cc2f2719e0cb3e1f6d66c9 not found: ID does not exist" Oct 08 09:44:41 crc kubenswrapper[4744]: I1008 09:44:41.465926 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" path="/var/lib/kubelet/pods/8c3a3397-3ed0-49fd-9ab4-4e6f81909a86/volumes" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.144167 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z"] Oct 08 09:45:00 crc kubenswrapper[4744]: E1008 09:45:00.145211 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" containerName="extract-content" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.145230 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" containerName="extract-content" Oct 08 09:45:00 crc kubenswrapper[4744]: E1008 09:45:00.145247 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" containerName="registry-server" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.145255 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" containerName="registry-server" Oct 08 09:45:00 crc kubenswrapper[4744]: E1008 09:45:00.145299 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" containerName="extract-utilities" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.145308 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" containerName="extract-utilities" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.145525 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c3a3397-3ed0-49fd-9ab4-4e6f81909a86" containerName="registry-server" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.146307 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.148660 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.148677 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.162337 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z"] Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.300071 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79b1044-6136-4858-93f3-6386a59653f0-secret-volume\") pod \"collect-profiles-29331945-8279z\" (UID: \"d79b1044-6136-4858-93f3-6386a59653f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.300132 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fb2l\" (UniqueName: \"kubernetes.io/projected/d79b1044-6136-4858-93f3-6386a59653f0-kube-api-access-4fb2l\") pod \"collect-profiles-29331945-8279z\" (UID: \"d79b1044-6136-4858-93f3-6386a59653f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.300891 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79b1044-6136-4858-93f3-6386a59653f0-config-volume\") pod \"collect-profiles-29331945-8279z\" (UID: \"d79b1044-6136-4858-93f3-6386a59653f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.402898 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79b1044-6136-4858-93f3-6386a59653f0-secret-volume\") pod \"collect-profiles-29331945-8279z\" (UID: \"d79b1044-6136-4858-93f3-6386a59653f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.403226 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fb2l\" (UniqueName: \"kubernetes.io/projected/d79b1044-6136-4858-93f3-6386a59653f0-kube-api-access-4fb2l\") pod \"collect-profiles-29331945-8279z\" (UID: \"d79b1044-6136-4858-93f3-6386a59653f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.403773 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79b1044-6136-4858-93f3-6386a59653f0-config-volume\") pod \"collect-profiles-29331945-8279z\" (UID: \"d79b1044-6136-4858-93f3-6386a59653f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.404658 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79b1044-6136-4858-93f3-6386a59653f0-config-volume\") pod \"collect-profiles-29331945-8279z\" (UID: \"d79b1044-6136-4858-93f3-6386a59653f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.416258 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79b1044-6136-4858-93f3-6386a59653f0-secret-volume\") pod \"collect-profiles-29331945-8279z\" (UID: \"d79b1044-6136-4858-93f3-6386a59653f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.419726 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fb2l\" (UniqueName: \"kubernetes.io/projected/d79b1044-6136-4858-93f3-6386a59653f0-kube-api-access-4fb2l\") pod \"collect-profiles-29331945-8279z\" (UID: \"d79b1044-6136-4858-93f3-6386a59653f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.517655 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:00 crc kubenswrapper[4744]: I1008 09:45:00.969486 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z"] Oct 08 09:45:01 crc kubenswrapper[4744]: I1008 09:45:01.880202 4744 generic.go:334] "Generic (PLEG): container finished" podID="d79b1044-6136-4858-93f3-6386a59653f0" containerID="2db52d6c8e769a1197b7e6367f657bdbef3be33119fc78af04557521a5c5a858" exitCode=0 Oct 08 09:45:01 crc kubenswrapper[4744]: I1008 09:45:01.880517 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" event={"ID":"d79b1044-6136-4858-93f3-6386a59653f0","Type":"ContainerDied","Data":"2db52d6c8e769a1197b7e6367f657bdbef3be33119fc78af04557521a5c5a858"} Oct 08 09:45:01 crc kubenswrapper[4744]: I1008 09:45:01.880746 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" event={"ID":"d79b1044-6136-4858-93f3-6386a59653f0","Type":"ContainerStarted","Data":"fdd403a64f2765900928152f033462ba4a60b8678831b3f76e6b76f906f99ba6"} Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.200972 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.362842 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fb2l\" (UniqueName: \"kubernetes.io/projected/d79b1044-6136-4858-93f3-6386a59653f0-kube-api-access-4fb2l\") pod \"d79b1044-6136-4858-93f3-6386a59653f0\" (UID: \"d79b1044-6136-4858-93f3-6386a59653f0\") " Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.363133 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79b1044-6136-4858-93f3-6386a59653f0-config-volume\") pod \"d79b1044-6136-4858-93f3-6386a59653f0\" (UID: \"d79b1044-6136-4858-93f3-6386a59653f0\") " Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.363193 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79b1044-6136-4858-93f3-6386a59653f0-secret-volume\") pod \"d79b1044-6136-4858-93f3-6386a59653f0\" (UID: \"d79b1044-6136-4858-93f3-6386a59653f0\") " Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.363859 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d79b1044-6136-4858-93f3-6386a59653f0-config-volume" (OuterVolumeSpecName: "config-volume") pod "d79b1044-6136-4858-93f3-6386a59653f0" (UID: "d79b1044-6136-4858-93f3-6386a59653f0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.368581 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d79b1044-6136-4858-93f3-6386a59653f0-kube-api-access-4fb2l" (OuterVolumeSpecName: "kube-api-access-4fb2l") pod "d79b1044-6136-4858-93f3-6386a59653f0" (UID: "d79b1044-6136-4858-93f3-6386a59653f0"). InnerVolumeSpecName "kube-api-access-4fb2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.368878 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79b1044-6136-4858-93f3-6386a59653f0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d79b1044-6136-4858-93f3-6386a59653f0" (UID: "d79b1044-6136-4858-93f3-6386a59653f0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.465238 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79b1044-6136-4858-93f3-6386a59653f0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.465274 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79b1044-6136-4858-93f3-6386a59653f0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.465288 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fb2l\" (UniqueName: \"kubernetes.io/projected/d79b1044-6136-4858-93f3-6386a59653f0-kube-api-access-4fb2l\") on node \"crc\" DevicePath \"\"" Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.898087 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" event={"ID":"d79b1044-6136-4858-93f3-6386a59653f0","Type":"ContainerDied","Data":"fdd403a64f2765900928152f033462ba4a60b8678831b3f76e6b76f906f99ba6"} Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.898129 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdd403a64f2765900928152f033462ba4a60b8678831b3f76e6b76f906f99ba6" Oct 08 09:45:03 crc kubenswrapper[4744]: I1008 09:45:03.898115 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z" Oct 08 09:45:04 crc kubenswrapper[4744]: I1008 09:45:04.276087 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv"] Oct 08 09:45:04 crc kubenswrapper[4744]: I1008 09:45:04.293927 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331900-b46lv"] Oct 08 09:45:05 crc kubenswrapper[4744]: I1008 09:45:05.467639 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bacbc855-fe99-4250-9da3-b64e95ac7c67" path="/var/lib/kubelet/pods/bacbc855-fe99-4250-9da3-b64e95ac7c67/volumes" Oct 08 09:45:19 crc kubenswrapper[4744]: I1008 09:45:19.690462 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:45:19 crc kubenswrapper[4744]: I1008 09:45:19.690983 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:45:49 crc kubenswrapper[4744]: I1008 09:45:49.690207 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:45:49 crc kubenswrapper[4744]: I1008 09:45:49.690863 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:45:55 crc kubenswrapper[4744]: I1008 09:45:55.249953 4744 scope.go:117] "RemoveContainer" containerID="ac4f0742774ae00345863f7dce979932ec68a8860401bf8a5a775c35facb8fab" Oct 08 09:46:19 crc kubenswrapper[4744]: I1008 09:46:19.690146 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:46:19 crc kubenswrapper[4744]: I1008 09:46:19.690811 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:46:19 crc kubenswrapper[4744]: I1008 09:46:19.690864 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:46:19 crc kubenswrapper[4744]: I1008 09:46:19.691749 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"edfca7ab5f128667cb0425a5ce4b92f849206f6e3061358fe9d17c2c761c3416"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 09:46:19 crc kubenswrapper[4744]: I1008 09:46:19.691824 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://edfca7ab5f128667cb0425a5ce4b92f849206f6e3061358fe9d17c2c761c3416" gracePeriod=600 Oct 08 09:46:20 crc kubenswrapper[4744]: I1008 09:46:20.623454 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="edfca7ab5f128667cb0425a5ce4b92f849206f6e3061358fe9d17c2c761c3416" exitCode=0 Oct 08 09:46:20 crc kubenswrapper[4744]: I1008 09:46:20.623538 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"edfca7ab5f128667cb0425a5ce4b92f849206f6e3061358fe9d17c2c761c3416"} Oct 08 09:46:20 crc kubenswrapper[4744]: I1008 09:46:20.624364 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af"} Oct 08 09:46:20 crc kubenswrapper[4744]: I1008 09:46:20.624431 4744 scope.go:117] "RemoveContainer" containerID="0d3f5ed4107e3668ce06d205a8c2d232d085cae862537b233fb5195e7ee25bcb" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.207662 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q4jrt"] Oct 08 09:46:25 crc kubenswrapper[4744]: E1008 09:46:25.208808 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d79b1044-6136-4858-93f3-6386a59653f0" containerName="collect-profiles" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.208827 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d79b1044-6136-4858-93f3-6386a59653f0" containerName="collect-profiles" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.209066 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d79b1044-6136-4858-93f3-6386a59653f0" containerName="collect-profiles" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.210626 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.233120 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4jrt"] Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.258309 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6949c5ba-2d03-4684-98c5-d8b755ff5820-utilities\") pod \"redhat-marketplace-q4jrt\" (UID: \"6949c5ba-2d03-4684-98c5-d8b755ff5820\") " pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.258352 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6949c5ba-2d03-4684-98c5-d8b755ff5820-catalog-content\") pod \"redhat-marketplace-q4jrt\" (UID: \"6949c5ba-2d03-4684-98c5-d8b755ff5820\") " pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.258423 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl9dn\" (UniqueName: \"kubernetes.io/projected/6949c5ba-2d03-4684-98c5-d8b755ff5820-kube-api-access-hl9dn\") pod \"redhat-marketplace-q4jrt\" (UID: \"6949c5ba-2d03-4684-98c5-d8b755ff5820\") " pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.360848 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6949c5ba-2d03-4684-98c5-d8b755ff5820-utilities\") pod \"redhat-marketplace-q4jrt\" (UID: \"6949c5ba-2d03-4684-98c5-d8b755ff5820\") " pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.360906 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6949c5ba-2d03-4684-98c5-d8b755ff5820-catalog-content\") pod \"redhat-marketplace-q4jrt\" (UID: \"6949c5ba-2d03-4684-98c5-d8b755ff5820\") " pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.360960 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl9dn\" (UniqueName: \"kubernetes.io/projected/6949c5ba-2d03-4684-98c5-d8b755ff5820-kube-api-access-hl9dn\") pod \"redhat-marketplace-q4jrt\" (UID: \"6949c5ba-2d03-4684-98c5-d8b755ff5820\") " pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.361455 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6949c5ba-2d03-4684-98c5-d8b755ff5820-utilities\") pod \"redhat-marketplace-q4jrt\" (UID: \"6949c5ba-2d03-4684-98c5-d8b755ff5820\") " pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.361534 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6949c5ba-2d03-4684-98c5-d8b755ff5820-catalog-content\") pod \"redhat-marketplace-q4jrt\" (UID: \"6949c5ba-2d03-4684-98c5-d8b755ff5820\") " pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.378527 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl9dn\" (UniqueName: \"kubernetes.io/projected/6949c5ba-2d03-4684-98c5-d8b755ff5820-kube-api-access-hl9dn\") pod \"redhat-marketplace-q4jrt\" (UID: \"6949c5ba-2d03-4684-98c5-d8b755ff5820\") " pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:25 crc kubenswrapper[4744]: I1008 09:46:25.530557 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:26 crc kubenswrapper[4744]: I1008 09:46:26.001613 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4jrt"] Oct 08 09:46:26 crc kubenswrapper[4744]: I1008 09:46:26.698911 4744 generic.go:334] "Generic (PLEG): container finished" podID="6949c5ba-2d03-4684-98c5-d8b755ff5820" containerID="4e6ecab2a356cd14b79a637edeb351088b8c86b2d1318318c4925d0af18f43e1" exitCode=0 Oct 08 09:46:26 crc kubenswrapper[4744]: I1008 09:46:26.699010 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jrt" event={"ID":"6949c5ba-2d03-4684-98c5-d8b755ff5820","Type":"ContainerDied","Data":"4e6ecab2a356cd14b79a637edeb351088b8c86b2d1318318c4925d0af18f43e1"} Oct 08 09:46:26 crc kubenswrapper[4744]: I1008 09:46:26.699196 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jrt" event={"ID":"6949c5ba-2d03-4684-98c5-d8b755ff5820","Type":"ContainerStarted","Data":"8835d12444b45d060bb9eab73182adb87154d47e3621c4032e8ce06403e4a8a4"} Oct 08 09:46:27 crc kubenswrapper[4744]: I1008 09:46:27.712877 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jrt" event={"ID":"6949c5ba-2d03-4684-98c5-d8b755ff5820","Type":"ContainerStarted","Data":"9b15ad055f91e232e7ef92bdc1f64b783e43e96ff73c283750a33352994fc680"} Oct 08 09:46:28 crc kubenswrapper[4744]: I1008 09:46:28.726706 4744 generic.go:334] "Generic (PLEG): container finished" podID="6949c5ba-2d03-4684-98c5-d8b755ff5820" containerID="9b15ad055f91e232e7ef92bdc1f64b783e43e96ff73c283750a33352994fc680" exitCode=0 Oct 08 09:46:28 crc kubenswrapper[4744]: I1008 09:46:28.726771 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jrt" event={"ID":"6949c5ba-2d03-4684-98c5-d8b755ff5820","Type":"ContainerDied","Data":"9b15ad055f91e232e7ef92bdc1f64b783e43e96ff73c283750a33352994fc680"} Oct 08 09:46:29 crc kubenswrapper[4744]: I1008 09:46:29.736497 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jrt" event={"ID":"6949c5ba-2d03-4684-98c5-d8b755ff5820","Type":"ContainerStarted","Data":"1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb"} Oct 08 09:46:29 crc kubenswrapper[4744]: I1008 09:46:29.762678 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q4jrt" podStartSLOduration=2.342200026 podStartE2EDuration="4.762662102s" podCreationTimestamp="2025-10-08 09:46:25 +0000 UTC" firstStartedPulling="2025-10-08 09:46:26.702506774 +0000 UTC m=+2081.950152013" lastFinishedPulling="2025-10-08 09:46:29.12296885 +0000 UTC m=+2084.370614089" observedRunningTime="2025-10-08 09:46:29.762522498 +0000 UTC m=+2085.010167747" watchObservedRunningTime="2025-10-08 09:46:29.762662102 +0000 UTC m=+2085.010307341" Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.006361 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xgzsn"] Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.008664 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.032046 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xgzsn"] Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.071671 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45rnt\" (UniqueName: \"kubernetes.io/projected/54f9a038-7cea-42f3-8d2c-6f5756abb98c-kube-api-access-45rnt\") pod \"community-operators-xgzsn\" (UID: \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\") " pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.072124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54f9a038-7cea-42f3-8d2c-6f5756abb98c-catalog-content\") pod \"community-operators-xgzsn\" (UID: \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\") " pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.072244 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54f9a038-7cea-42f3-8d2c-6f5756abb98c-utilities\") pod \"community-operators-xgzsn\" (UID: \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\") " pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.173656 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45rnt\" (UniqueName: \"kubernetes.io/projected/54f9a038-7cea-42f3-8d2c-6f5756abb98c-kube-api-access-45rnt\") pod \"community-operators-xgzsn\" (UID: \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\") " pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.173929 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54f9a038-7cea-42f3-8d2c-6f5756abb98c-catalog-content\") pod \"community-operators-xgzsn\" (UID: \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\") " pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.174061 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54f9a038-7cea-42f3-8d2c-6f5756abb98c-utilities\") pod \"community-operators-xgzsn\" (UID: \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\") " pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.174792 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54f9a038-7cea-42f3-8d2c-6f5756abb98c-utilities\") pod \"community-operators-xgzsn\" (UID: \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\") " pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.174929 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54f9a038-7cea-42f3-8d2c-6f5756abb98c-catalog-content\") pod \"community-operators-xgzsn\" (UID: \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\") " pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.202146 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45rnt\" (UniqueName: \"kubernetes.io/projected/54f9a038-7cea-42f3-8d2c-6f5756abb98c-kube-api-access-45rnt\") pod \"community-operators-xgzsn\" (UID: \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\") " pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.328520 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:31 crc kubenswrapper[4744]: I1008 09:46:31.934757 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xgzsn"] Oct 08 09:46:31 crc kubenswrapper[4744]: W1008 09:46:31.938978 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54f9a038_7cea_42f3_8d2c_6f5756abb98c.slice/crio-fbbeff7377dd80f99c011fb817f68c447ab514b639a9846c6093d41117506fc0 WatchSource:0}: Error finding container fbbeff7377dd80f99c011fb817f68c447ab514b639a9846c6093d41117506fc0: Status 404 returned error can't find the container with id fbbeff7377dd80f99c011fb817f68c447ab514b639a9846c6093d41117506fc0 Oct 08 09:46:32 crc kubenswrapper[4744]: I1008 09:46:32.767789 4744 generic.go:334] "Generic (PLEG): container finished" podID="54f9a038-7cea-42f3-8d2c-6f5756abb98c" containerID="9587b17bf96fd2729fc5296da9a0c0e5369a801cad117c3a94a7cc46720b694b" exitCode=0 Oct 08 09:46:32 crc kubenswrapper[4744]: I1008 09:46:32.767832 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzsn" event={"ID":"54f9a038-7cea-42f3-8d2c-6f5756abb98c","Type":"ContainerDied","Data":"9587b17bf96fd2729fc5296da9a0c0e5369a801cad117c3a94a7cc46720b694b"} Oct 08 09:46:32 crc kubenswrapper[4744]: I1008 09:46:32.767860 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzsn" event={"ID":"54f9a038-7cea-42f3-8d2c-6f5756abb98c","Type":"ContainerStarted","Data":"fbbeff7377dd80f99c011fb817f68c447ab514b639a9846c6093d41117506fc0"} Oct 08 09:46:33 crc kubenswrapper[4744]: I1008 09:46:33.781190 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzsn" event={"ID":"54f9a038-7cea-42f3-8d2c-6f5756abb98c","Type":"ContainerStarted","Data":"0714e1c8fb5a404b27340ace9a284ab0bc1e663c19516e4eaa5825e8e5acb196"} Oct 08 09:46:34 crc kubenswrapper[4744]: I1008 09:46:34.792787 4744 generic.go:334] "Generic (PLEG): container finished" podID="54f9a038-7cea-42f3-8d2c-6f5756abb98c" containerID="0714e1c8fb5a404b27340ace9a284ab0bc1e663c19516e4eaa5825e8e5acb196" exitCode=0 Oct 08 09:46:34 crc kubenswrapper[4744]: I1008 09:46:34.792914 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzsn" event={"ID":"54f9a038-7cea-42f3-8d2c-6f5756abb98c","Type":"ContainerDied","Data":"0714e1c8fb5a404b27340ace9a284ab0bc1e663c19516e4eaa5825e8e5acb196"} Oct 08 09:46:35 crc kubenswrapper[4744]: I1008 09:46:35.532017 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:35 crc kubenswrapper[4744]: I1008 09:46:35.532660 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:35 crc kubenswrapper[4744]: I1008 09:46:35.582034 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:35 crc kubenswrapper[4744]: I1008 09:46:35.803775 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzsn" event={"ID":"54f9a038-7cea-42f3-8d2c-6f5756abb98c","Type":"ContainerStarted","Data":"e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5"} Oct 08 09:46:35 crc kubenswrapper[4744]: I1008 09:46:35.827125 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xgzsn" podStartSLOduration=3.370700951 podStartE2EDuration="5.827102606s" podCreationTimestamp="2025-10-08 09:46:30 +0000 UTC" firstStartedPulling="2025-10-08 09:46:32.769800916 +0000 UTC m=+2088.017446155" lastFinishedPulling="2025-10-08 09:46:35.226202571 +0000 UTC m=+2090.473847810" observedRunningTime="2025-10-08 09:46:35.823588458 +0000 UTC m=+2091.071233707" watchObservedRunningTime="2025-10-08 09:46:35.827102606 +0000 UTC m=+2091.074747845" Oct 08 09:46:35 crc kubenswrapper[4744]: I1008 09:46:35.867986 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:37 crc kubenswrapper[4744]: I1008 09:46:37.200282 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4jrt"] Oct 08 09:46:37 crc kubenswrapper[4744]: I1008 09:46:37.819516 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q4jrt" podUID="6949c5ba-2d03-4684-98c5-d8b755ff5820" containerName="registry-server" containerID="cri-o://1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb" gracePeriod=2 Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.292776 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.468020 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hl9dn\" (UniqueName: \"kubernetes.io/projected/6949c5ba-2d03-4684-98c5-d8b755ff5820-kube-api-access-hl9dn\") pod \"6949c5ba-2d03-4684-98c5-d8b755ff5820\" (UID: \"6949c5ba-2d03-4684-98c5-d8b755ff5820\") " Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.468279 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6949c5ba-2d03-4684-98c5-d8b755ff5820-catalog-content\") pod \"6949c5ba-2d03-4684-98c5-d8b755ff5820\" (UID: \"6949c5ba-2d03-4684-98c5-d8b755ff5820\") " Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.468319 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6949c5ba-2d03-4684-98c5-d8b755ff5820-utilities\") pod \"6949c5ba-2d03-4684-98c5-d8b755ff5820\" (UID: \"6949c5ba-2d03-4684-98c5-d8b755ff5820\") " Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.469608 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6949c5ba-2d03-4684-98c5-d8b755ff5820-utilities" (OuterVolumeSpecName: "utilities") pod "6949c5ba-2d03-4684-98c5-d8b755ff5820" (UID: "6949c5ba-2d03-4684-98c5-d8b755ff5820"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.476604 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6949c5ba-2d03-4684-98c5-d8b755ff5820-kube-api-access-hl9dn" (OuterVolumeSpecName: "kube-api-access-hl9dn") pod "6949c5ba-2d03-4684-98c5-d8b755ff5820" (UID: "6949c5ba-2d03-4684-98c5-d8b755ff5820"). InnerVolumeSpecName "kube-api-access-hl9dn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.488456 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6949c5ba-2d03-4684-98c5-d8b755ff5820-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6949c5ba-2d03-4684-98c5-d8b755ff5820" (UID: "6949c5ba-2d03-4684-98c5-d8b755ff5820"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.572794 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hl9dn\" (UniqueName: \"kubernetes.io/projected/6949c5ba-2d03-4684-98c5-d8b755ff5820-kube-api-access-hl9dn\") on node \"crc\" DevicePath \"\"" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.572934 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6949c5ba-2d03-4684-98c5-d8b755ff5820-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.572960 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6949c5ba-2d03-4684-98c5-d8b755ff5820-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.832443 4744 generic.go:334] "Generic (PLEG): container finished" podID="6949c5ba-2d03-4684-98c5-d8b755ff5820" containerID="1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb" exitCode=0 Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.832510 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jrt" event={"ID":"6949c5ba-2d03-4684-98c5-d8b755ff5820","Type":"ContainerDied","Data":"1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb"} Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.832539 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4jrt" event={"ID":"6949c5ba-2d03-4684-98c5-d8b755ff5820","Type":"ContainerDied","Data":"8835d12444b45d060bb9eab73182adb87154d47e3621c4032e8ce06403e4a8a4"} Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.832558 4744 scope.go:117] "RemoveContainer" containerID="1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.832582 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4jrt" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.855529 4744 scope.go:117] "RemoveContainer" containerID="9b15ad055f91e232e7ef92bdc1f64b783e43e96ff73c283750a33352994fc680" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.870544 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4jrt"] Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.878245 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4jrt"] Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.888606 4744 scope.go:117] "RemoveContainer" containerID="4e6ecab2a356cd14b79a637edeb351088b8c86b2d1318318c4925d0af18f43e1" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.940458 4744 scope.go:117] "RemoveContainer" containerID="1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb" Oct 08 09:46:38 crc kubenswrapper[4744]: E1008 09:46:38.940869 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb\": container with ID starting with 1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb not found: ID does not exist" containerID="1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.940901 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb"} err="failed to get container status \"1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb\": rpc error: code = NotFound desc = could not find container \"1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb\": container with ID starting with 1537c9fdf20e5dc19d6b66707092e80fe56ff81d25d1fc3aaebe522072604abb not found: ID does not exist" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.940925 4744 scope.go:117] "RemoveContainer" containerID="9b15ad055f91e232e7ef92bdc1f64b783e43e96ff73c283750a33352994fc680" Oct 08 09:46:38 crc kubenswrapper[4744]: E1008 09:46:38.941185 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b15ad055f91e232e7ef92bdc1f64b783e43e96ff73c283750a33352994fc680\": container with ID starting with 9b15ad055f91e232e7ef92bdc1f64b783e43e96ff73c283750a33352994fc680 not found: ID does not exist" containerID="9b15ad055f91e232e7ef92bdc1f64b783e43e96ff73c283750a33352994fc680" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.941231 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b15ad055f91e232e7ef92bdc1f64b783e43e96ff73c283750a33352994fc680"} err="failed to get container status \"9b15ad055f91e232e7ef92bdc1f64b783e43e96ff73c283750a33352994fc680\": rpc error: code = NotFound desc = could not find container \"9b15ad055f91e232e7ef92bdc1f64b783e43e96ff73c283750a33352994fc680\": container with ID starting with 9b15ad055f91e232e7ef92bdc1f64b783e43e96ff73c283750a33352994fc680 not found: ID does not exist" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.941261 4744 scope.go:117] "RemoveContainer" containerID="4e6ecab2a356cd14b79a637edeb351088b8c86b2d1318318c4925d0af18f43e1" Oct 08 09:46:38 crc kubenswrapper[4744]: E1008 09:46:38.942438 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e6ecab2a356cd14b79a637edeb351088b8c86b2d1318318c4925d0af18f43e1\": container with ID starting with 4e6ecab2a356cd14b79a637edeb351088b8c86b2d1318318c4925d0af18f43e1 not found: ID does not exist" containerID="4e6ecab2a356cd14b79a637edeb351088b8c86b2d1318318c4925d0af18f43e1" Oct 08 09:46:38 crc kubenswrapper[4744]: I1008 09:46:38.942523 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e6ecab2a356cd14b79a637edeb351088b8c86b2d1318318c4925d0af18f43e1"} err="failed to get container status \"4e6ecab2a356cd14b79a637edeb351088b8c86b2d1318318c4925d0af18f43e1\": rpc error: code = NotFound desc = could not find container \"4e6ecab2a356cd14b79a637edeb351088b8c86b2d1318318c4925d0af18f43e1\": container with ID starting with 4e6ecab2a356cd14b79a637edeb351088b8c86b2d1318318c4925d0af18f43e1 not found: ID does not exist" Oct 08 09:46:39 crc kubenswrapper[4744]: I1008 09:46:39.465939 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6949c5ba-2d03-4684-98c5-d8b755ff5820" path="/var/lib/kubelet/pods/6949c5ba-2d03-4684-98c5-d8b755ff5820/volumes" Oct 08 09:46:41 crc kubenswrapper[4744]: I1008 09:46:41.329270 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:41 crc kubenswrapper[4744]: I1008 09:46:41.329312 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:41 crc kubenswrapper[4744]: I1008 09:46:41.377083 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:41 crc kubenswrapper[4744]: I1008 09:46:41.905985 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:42 crc kubenswrapper[4744]: I1008 09:46:42.200346 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xgzsn"] Oct 08 09:46:43 crc kubenswrapper[4744]: I1008 09:46:43.872033 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xgzsn" podUID="54f9a038-7cea-42f3-8d2c-6f5756abb98c" containerName="registry-server" containerID="cri-o://e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5" gracePeriod=2 Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.326270 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.471703 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45rnt\" (UniqueName: \"kubernetes.io/projected/54f9a038-7cea-42f3-8d2c-6f5756abb98c-kube-api-access-45rnt\") pod \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\" (UID: \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\") " Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.471801 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54f9a038-7cea-42f3-8d2c-6f5756abb98c-catalog-content\") pod \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\" (UID: \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\") " Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.471851 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54f9a038-7cea-42f3-8d2c-6f5756abb98c-utilities\") pod \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\" (UID: \"54f9a038-7cea-42f3-8d2c-6f5756abb98c\") " Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.474591 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54f9a038-7cea-42f3-8d2c-6f5756abb98c-utilities" (OuterVolumeSpecName: "utilities") pod "54f9a038-7cea-42f3-8d2c-6f5756abb98c" (UID: "54f9a038-7cea-42f3-8d2c-6f5756abb98c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.479063 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54f9a038-7cea-42f3-8d2c-6f5756abb98c-kube-api-access-45rnt" (OuterVolumeSpecName: "kube-api-access-45rnt") pod "54f9a038-7cea-42f3-8d2c-6f5756abb98c" (UID: "54f9a038-7cea-42f3-8d2c-6f5756abb98c"). InnerVolumeSpecName "kube-api-access-45rnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.532297 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54f9a038-7cea-42f3-8d2c-6f5756abb98c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54f9a038-7cea-42f3-8d2c-6f5756abb98c" (UID: "54f9a038-7cea-42f3-8d2c-6f5756abb98c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.573997 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45rnt\" (UniqueName: \"kubernetes.io/projected/54f9a038-7cea-42f3-8d2c-6f5756abb98c-kube-api-access-45rnt\") on node \"crc\" DevicePath \"\"" Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.574043 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54f9a038-7cea-42f3-8d2c-6f5756abb98c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.574056 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54f9a038-7cea-42f3-8d2c-6f5756abb98c-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.898154 4744 generic.go:334] "Generic (PLEG): container finished" podID="54f9a038-7cea-42f3-8d2c-6f5756abb98c" containerID="e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5" exitCode=0 Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.898249 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzsn" event={"ID":"54f9a038-7cea-42f3-8d2c-6f5756abb98c","Type":"ContainerDied","Data":"e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5"} Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.898524 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzsn" event={"ID":"54f9a038-7cea-42f3-8d2c-6f5756abb98c","Type":"ContainerDied","Data":"fbbeff7377dd80f99c011fb817f68c447ab514b639a9846c6093d41117506fc0"} Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.898554 4744 scope.go:117] "RemoveContainer" containerID="e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5" Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.898320 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgzsn" Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.931217 4744 scope.go:117] "RemoveContainer" containerID="0714e1c8fb5a404b27340ace9a284ab0bc1e663c19516e4eaa5825e8e5acb196" Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.939426 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xgzsn"] Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.947020 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xgzsn"] Oct 08 09:46:44 crc kubenswrapper[4744]: I1008 09:46:44.959456 4744 scope.go:117] "RemoveContainer" containerID="9587b17bf96fd2729fc5296da9a0c0e5369a801cad117c3a94a7cc46720b694b" Oct 08 09:46:45 crc kubenswrapper[4744]: I1008 09:46:45.003437 4744 scope.go:117] "RemoveContainer" containerID="e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5" Oct 08 09:46:45 crc kubenswrapper[4744]: E1008 09:46:45.004042 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5\": container with ID starting with e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5 not found: ID does not exist" containerID="e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5" Oct 08 09:46:45 crc kubenswrapper[4744]: I1008 09:46:45.004126 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5"} err="failed to get container status \"e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5\": rpc error: code = NotFound desc = could not find container \"e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5\": container with ID starting with e6bf1b7e7a2c90967c4a96a7bed9399e9facc31136e69fc2d6e4fa8bb92e34d5 not found: ID does not exist" Oct 08 09:46:45 crc kubenswrapper[4744]: I1008 09:46:45.004172 4744 scope.go:117] "RemoveContainer" containerID="0714e1c8fb5a404b27340ace9a284ab0bc1e663c19516e4eaa5825e8e5acb196" Oct 08 09:46:45 crc kubenswrapper[4744]: E1008 09:46:45.004622 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0714e1c8fb5a404b27340ace9a284ab0bc1e663c19516e4eaa5825e8e5acb196\": container with ID starting with 0714e1c8fb5a404b27340ace9a284ab0bc1e663c19516e4eaa5825e8e5acb196 not found: ID does not exist" containerID="0714e1c8fb5a404b27340ace9a284ab0bc1e663c19516e4eaa5825e8e5acb196" Oct 08 09:46:45 crc kubenswrapper[4744]: I1008 09:46:45.004661 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0714e1c8fb5a404b27340ace9a284ab0bc1e663c19516e4eaa5825e8e5acb196"} err="failed to get container status \"0714e1c8fb5a404b27340ace9a284ab0bc1e663c19516e4eaa5825e8e5acb196\": rpc error: code = NotFound desc = could not find container \"0714e1c8fb5a404b27340ace9a284ab0bc1e663c19516e4eaa5825e8e5acb196\": container with ID starting with 0714e1c8fb5a404b27340ace9a284ab0bc1e663c19516e4eaa5825e8e5acb196 not found: ID does not exist" Oct 08 09:46:45 crc kubenswrapper[4744]: I1008 09:46:45.004688 4744 scope.go:117] "RemoveContainer" containerID="9587b17bf96fd2729fc5296da9a0c0e5369a801cad117c3a94a7cc46720b694b" Oct 08 09:46:45 crc kubenswrapper[4744]: E1008 09:46:45.005088 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9587b17bf96fd2729fc5296da9a0c0e5369a801cad117c3a94a7cc46720b694b\": container with ID starting with 9587b17bf96fd2729fc5296da9a0c0e5369a801cad117c3a94a7cc46720b694b not found: ID does not exist" containerID="9587b17bf96fd2729fc5296da9a0c0e5369a801cad117c3a94a7cc46720b694b" Oct 08 09:46:45 crc kubenswrapper[4744]: I1008 09:46:45.005148 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9587b17bf96fd2729fc5296da9a0c0e5369a801cad117c3a94a7cc46720b694b"} err="failed to get container status \"9587b17bf96fd2729fc5296da9a0c0e5369a801cad117c3a94a7cc46720b694b\": rpc error: code = NotFound desc = could not find container \"9587b17bf96fd2729fc5296da9a0c0e5369a801cad117c3a94a7cc46720b694b\": container with ID starting with 9587b17bf96fd2729fc5296da9a0c0e5369a801cad117c3a94a7cc46720b694b not found: ID does not exist" Oct 08 09:46:45 crc kubenswrapper[4744]: I1008 09:46:45.466298 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54f9a038-7cea-42f3-8d2c-6f5756abb98c" path="/var/lib/kubelet/pods/54f9a038-7cea-42f3-8d2c-6f5756abb98c/volumes" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.336565 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kf8r7"] Oct 08 09:47:27 crc kubenswrapper[4744]: E1008 09:47:27.339209 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54f9a038-7cea-42f3-8d2c-6f5756abb98c" containerName="registry-server" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.339312 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="54f9a038-7cea-42f3-8d2c-6f5756abb98c" containerName="registry-server" Oct 08 09:47:27 crc kubenswrapper[4744]: E1008 09:47:27.339441 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6949c5ba-2d03-4684-98c5-d8b755ff5820" containerName="extract-content" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.339813 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6949c5ba-2d03-4684-98c5-d8b755ff5820" containerName="extract-content" Oct 08 09:47:27 crc kubenswrapper[4744]: E1008 09:47:27.339891 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54f9a038-7cea-42f3-8d2c-6f5756abb98c" containerName="extract-utilities" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.340003 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="54f9a038-7cea-42f3-8d2c-6f5756abb98c" containerName="extract-utilities" Oct 08 09:47:27 crc kubenswrapper[4744]: E1008 09:47:27.340122 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6949c5ba-2d03-4684-98c5-d8b755ff5820" containerName="extract-utilities" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.340214 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6949c5ba-2d03-4684-98c5-d8b755ff5820" containerName="extract-utilities" Oct 08 09:47:27 crc kubenswrapper[4744]: E1008 09:47:27.340312 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54f9a038-7cea-42f3-8d2c-6f5756abb98c" containerName="extract-content" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.340423 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="54f9a038-7cea-42f3-8d2c-6f5756abb98c" containerName="extract-content" Oct 08 09:47:27 crc kubenswrapper[4744]: E1008 09:47:27.340500 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6949c5ba-2d03-4684-98c5-d8b755ff5820" containerName="registry-server" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.340666 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6949c5ba-2d03-4684-98c5-d8b755ff5820" containerName="registry-server" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.341016 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6949c5ba-2d03-4684-98c5-d8b755ff5820" containerName="registry-server" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.341124 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="54f9a038-7cea-42f3-8d2c-6f5756abb98c" containerName="registry-server" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.342855 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.408702 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kf8r7"] Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.424197 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/154a234a-1453-4689-9afe-00c47172bd87-utilities\") pod \"certified-operators-kf8r7\" (UID: \"154a234a-1453-4689-9afe-00c47172bd87\") " pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.424320 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65chh\" (UniqueName: \"kubernetes.io/projected/154a234a-1453-4689-9afe-00c47172bd87-kube-api-access-65chh\") pod \"certified-operators-kf8r7\" (UID: \"154a234a-1453-4689-9afe-00c47172bd87\") " pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.424444 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/154a234a-1453-4689-9afe-00c47172bd87-catalog-content\") pod \"certified-operators-kf8r7\" (UID: \"154a234a-1453-4689-9afe-00c47172bd87\") " pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.525543 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/154a234a-1453-4689-9afe-00c47172bd87-utilities\") pod \"certified-operators-kf8r7\" (UID: \"154a234a-1453-4689-9afe-00c47172bd87\") " pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.525606 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65chh\" (UniqueName: \"kubernetes.io/projected/154a234a-1453-4689-9afe-00c47172bd87-kube-api-access-65chh\") pod \"certified-operators-kf8r7\" (UID: \"154a234a-1453-4689-9afe-00c47172bd87\") " pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.525646 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/154a234a-1453-4689-9afe-00c47172bd87-catalog-content\") pod \"certified-operators-kf8r7\" (UID: \"154a234a-1453-4689-9afe-00c47172bd87\") " pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.526126 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/154a234a-1453-4689-9afe-00c47172bd87-utilities\") pod \"certified-operators-kf8r7\" (UID: \"154a234a-1453-4689-9afe-00c47172bd87\") " pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.526364 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/154a234a-1453-4689-9afe-00c47172bd87-catalog-content\") pod \"certified-operators-kf8r7\" (UID: \"154a234a-1453-4689-9afe-00c47172bd87\") " pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.548503 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65chh\" (UniqueName: \"kubernetes.io/projected/154a234a-1453-4689-9afe-00c47172bd87-kube-api-access-65chh\") pod \"certified-operators-kf8r7\" (UID: \"154a234a-1453-4689-9afe-00c47172bd87\") " pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:27 crc kubenswrapper[4744]: I1008 09:47:27.695525 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:28 crc kubenswrapper[4744]: I1008 09:47:28.230274 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kf8r7"] Oct 08 09:47:28 crc kubenswrapper[4744]: I1008 09:47:28.333760 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf8r7" event={"ID":"154a234a-1453-4689-9afe-00c47172bd87","Type":"ContainerStarted","Data":"fcfab7152f41d2e4f3532d642bd4e1ab82c7c8eff8616f107103321b192426ce"} Oct 08 09:47:29 crc kubenswrapper[4744]: I1008 09:47:29.349981 4744 generic.go:334] "Generic (PLEG): container finished" podID="154a234a-1453-4689-9afe-00c47172bd87" containerID="ae54bee9e483720e48891c1ab3de14273dbfe975609e075770b8976e6da89136" exitCode=0 Oct 08 09:47:29 crc kubenswrapper[4744]: I1008 09:47:29.350143 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf8r7" event={"ID":"154a234a-1453-4689-9afe-00c47172bd87","Type":"ContainerDied","Data":"ae54bee9e483720e48891c1ab3de14273dbfe975609e075770b8976e6da89136"} Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.676011 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.683803 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.709254 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-t4fcz"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.716342 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4ldnv"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.731057 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.740055 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4jwsc"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.746479 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.754974 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.761829 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.787418 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.791912 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.798720 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.805342 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.812540 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-4jwsc"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.818281 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-q28jc"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.825416 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6t8ms"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.832716 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-llqtn"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.840695 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-687kh"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.848578 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mnhk4"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.854397 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-48p2p"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.862007 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-9xgw7"] Oct 08 09:47:30 crc kubenswrapper[4744]: I1008 09:47:30.867005 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-q5zxg"] Oct 08 09:47:31 crc kubenswrapper[4744]: I1008 09:47:31.368685 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf8r7" event={"ID":"154a234a-1453-4689-9afe-00c47172bd87","Type":"ContainerStarted","Data":"5106304e01d8706c6d7ccf034b505634be3e34d6d59db81e3e1635c9c1fd2754"} Oct 08 09:47:31 crc kubenswrapper[4744]: I1008 09:47:31.463892 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c75eb05-f1da-4db0-87b6-7432e97f244c" path="/var/lib/kubelet/pods/0c75eb05-f1da-4db0-87b6-7432e97f244c/volumes" Oct 08 09:47:31 crc kubenswrapper[4744]: I1008 09:47:31.464832 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26f17329-55b8-4de4-82ee-98775dafa32a" path="/var/lib/kubelet/pods/26f17329-55b8-4de4-82ee-98775dafa32a/volumes" Oct 08 09:47:31 crc kubenswrapper[4744]: I1008 09:47:31.465526 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a036815-b901-4cd0-8ed9-b30c6c63be44" path="/var/lib/kubelet/pods/3a036815-b901-4cd0-8ed9-b30c6c63be44/volumes" Oct 08 09:47:31 crc kubenswrapper[4744]: I1008 09:47:31.466131 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="606f778d-ee4d-4e8a-8cfd-d2dc86874016" path="/var/lib/kubelet/pods/606f778d-ee4d-4e8a-8cfd-d2dc86874016/volumes" Oct 08 09:47:31 crc kubenswrapper[4744]: I1008 09:47:31.467358 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402ae4a-8d4f-4011-9a28-16b8b8303ba2" path="/var/lib/kubelet/pods/6402ae4a-8d4f-4011-9a28-16b8b8303ba2/volumes" Oct 08 09:47:31 crc kubenswrapper[4744]: I1008 09:47:31.467955 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ae696f3-08f6-4219-8b34-eb9afe223239" path="/var/lib/kubelet/pods/7ae696f3-08f6-4219-8b34-eb9afe223239/volumes" Oct 08 09:47:31 crc kubenswrapper[4744]: I1008 09:47:31.468524 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f69541a-2daf-4e28-b1a1-1811c5476652" path="/var/lib/kubelet/pods/7f69541a-2daf-4e28-b1a1-1811c5476652/volumes" Oct 08 09:47:31 crc kubenswrapper[4744]: I1008 09:47:31.469903 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9511d654-cff2-45ce-9b6a-a3d98744e0a6" path="/var/lib/kubelet/pods/9511d654-cff2-45ce-9b6a-a3d98744e0a6/volumes" Oct 08 09:47:31 crc kubenswrapper[4744]: I1008 09:47:31.474531 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a08d7867-ad28-4079-a99d-f6db36d9a323" path="/var/lib/kubelet/pods/a08d7867-ad28-4079-a99d-f6db36d9a323/volumes" Oct 08 09:47:31 crc kubenswrapper[4744]: I1008 09:47:31.477074 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e" path="/var/lib/kubelet/pods/c9c3e8f2-b1b0-4f64-8bc0-e880af536a4e/volumes" Oct 08 09:47:31 crc kubenswrapper[4744]: I1008 09:47:31.479064 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4b8727b-2bfd-4e9d-839d-8c7e69925c9b" path="/var/lib/kubelet/pods/e4b8727b-2bfd-4e9d-839d-8c7e69925c9b/volumes" Oct 08 09:47:32 crc kubenswrapper[4744]: I1008 09:47:32.380704 4744 generic.go:334] "Generic (PLEG): container finished" podID="154a234a-1453-4689-9afe-00c47172bd87" containerID="5106304e01d8706c6d7ccf034b505634be3e34d6d59db81e3e1635c9c1fd2754" exitCode=0 Oct 08 09:47:32 crc kubenswrapper[4744]: I1008 09:47:32.380840 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf8r7" event={"ID":"154a234a-1453-4689-9afe-00c47172bd87","Type":"ContainerDied","Data":"5106304e01d8706c6d7ccf034b505634be3e34d6d59db81e3e1635c9c1fd2754"} Oct 08 09:47:33 crc kubenswrapper[4744]: I1008 09:47:33.395534 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf8r7" event={"ID":"154a234a-1453-4689-9afe-00c47172bd87","Type":"ContainerStarted","Data":"fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff"} Oct 08 09:47:33 crc kubenswrapper[4744]: I1008 09:47:33.425623 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kf8r7" podStartSLOduration=2.996112112 podStartE2EDuration="6.425593281s" podCreationTimestamp="2025-10-08 09:47:27 +0000 UTC" firstStartedPulling="2025-10-08 09:47:29.353305424 +0000 UTC m=+2144.600950663" lastFinishedPulling="2025-10-08 09:47:32.782786593 +0000 UTC m=+2148.030431832" observedRunningTime="2025-10-08 09:47:33.416708014 +0000 UTC m=+2148.664353283" watchObservedRunningTime="2025-10-08 09:47:33.425593281 +0000 UTC m=+2148.673238520" Oct 08 09:47:37 crc kubenswrapper[4744]: I1008 09:47:37.696436 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:37 crc kubenswrapper[4744]: I1008 09:47:37.698965 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:37 crc kubenswrapper[4744]: I1008 09:47:37.771125 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:38 crc kubenswrapper[4744]: I1008 09:47:38.502995 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:38 crc kubenswrapper[4744]: I1008 09:47:38.563852 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kf8r7"] Oct 08 09:47:40 crc kubenswrapper[4744]: I1008 09:47:40.465234 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kf8r7" podUID="154a234a-1453-4689-9afe-00c47172bd87" containerName="registry-server" containerID="cri-o://fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff" gracePeriod=2 Oct 08 09:47:40 crc kubenswrapper[4744]: I1008 09:47:40.989884 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.053683 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/154a234a-1453-4689-9afe-00c47172bd87-utilities\") pod \"154a234a-1453-4689-9afe-00c47172bd87\" (UID: \"154a234a-1453-4689-9afe-00c47172bd87\") " Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.053768 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65chh\" (UniqueName: \"kubernetes.io/projected/154a234a-1453-4689-9afe-00c47172bd87-kube-api-access-65chh\") pod \"154a234a-1453-4689-9afe-00c47172bd87\" (UID: \"154a234a-1453-4689-9afe-00c47172bd87\") " Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.053872 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/154a234a-1453-4689-9afe-00c47172bd87-catalog-content\") pod \"154a234a-1453-4689-9afe-00c47172bd87\" (UID: \"154a234a-1453-4689-9afe-00c47172bd87\") " Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.056201 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/154a234a-1453-4689-9afe-00c47172bd87-utilities" (OuterVolumeSpecName: "utilities") pod "154a234a-1453-4689-9afe-00c47172bd87" (UID: "154a234a-1453-4689-9afe-00c47172bd87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.075185 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/154a234a-1453-4689-9afe-00c47172bd87-kube-api-access-65chh" (OuterVolumeSpecName: "kube-api-access-65chh") pod "154a234a-1453-4689-9afe-00c47172bd87" (UID: "154a234a-1453-4689-9afe-00c47172bd87"). InnerVolumeSpecName "kube-api-access-65chh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.102956 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/154a234a-1453-4689-9afe-00c47172bd87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "154a234a-1453-4689-9afe-00c47172bd87" (UID: "154a234a-1453-4689-9afe-00c47172bd87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.155454 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/154a234a-1453-4689-9afe-00c47172bd87-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.155488 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65chh\" (UniqueName: \"kubernetes.io/projected/154a234a-1453-4689-9afe-00c47172bd87-kube-api-access-65chh\") on node \"crc\" DevicePath \"\"" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.155500 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/154a234a-1453-4689-9afe-00c47172bd87-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.480416 4744 generic.go:334] "Generic (PLEG): container finished" podID="154a234a-1453-4689-9afe-00c47172bd87" containerID="fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff" exitCode=0 Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.480473 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kf8r7" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.480495 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf8r7" event={"ID":"154a234a-1453-4689-9afe-00c47172bd87","Type":"ContainerDied","Data":"fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff"} Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.481395 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kf8r7" event={"ID":"154a234a-1453-4689-9afe-00c47172bd87","Type":"ContainerDied","Data":"fcfab7152f41d2e4f3532d642bd4e1ab82c7c8eff8616f107103321b192426ce"} Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.481437 4744 scope.go:117] "RemoveContainer" containerID="fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.520998 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kf8r7"] Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.527864 4744 scope.go:117] "RemoveContainer" containerID="5106304e01d8706c6d7ccf034b505634be3e34d6d59db81e3e1635c9c1fd2754" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.533136 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kf8r7"] Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.557041 4744 scope.go:117] "RemoveContainer" containerID="ae54bee9e483720e48891c1ab3de14273dbfe975609e075770b8976e6da89136" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.599500 4744 scope.go:117] "RemoveContainer" containerID="fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff" Oct 08 09:47:41 crc kubenswrapper[4744]: E1008 09:47:41.600129 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff\": container with ID starting with fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff not found: ID does not exist" containerID="fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.600271 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff"} err="failed to get container status \"fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff\": rpc error: code = NotFound desc = could not find container \"fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff\": container with ID starting with fbcb6dd9948bfaa77eb74401dbaa399f19fe07f2e9409ba7306adf9956929aff not found: ID does not exist" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.600407 4744 scope.go:117] "RemoveContainer" containerID="5106304e01d8706c6d7ccf034b505634be3e34d6d59db81e3e1635c9c1fd2754" Oct 08 09:47:41 crc kubenswrapper[4744]: E1008 09:47:41.600814 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5106304e01d8706c6d7ccf034b505634be3e34d6d59db81e3e1635c9c1fd2754\": container with ID starting with 5106304e01d8706c6d7ccf034b505634be3e34d6d59db81e3e1635c9c1fd2754 not found: ID does not exist" containerID="5106304e01d8706c6d7ccf034b505634be3e34d6d59db81e3e1635c9c1fd2754" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.600865 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5106304e01d8706c6d7ccf034b505634be3e34d6d59db81e3e1635c9c1fd2754"} err="failed to get container status \"5106304e01d8706c6d7ccf034b505634be3e34d6d59db81e3e1635c9c1fd2754\": rpc error: code = NotFound desc = could not find container \"5106304e01d8706c6d7ccf034b505634be3e34d6d59db81e3e1635c9c1fd2754\": container with ID starting with 5106304e01d8706c6d7ccf034b505634be3e34d6d59db81e3e1635c9c1fd2754 not found: ID does not exist" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.600896 4744 scope.go:117] "RemoveContainer" containerID="ae54bee9e483720e48891c1ab3de14273dbfe975609e075770b8976e6da89136" Oct 08 09:47:41 crc kubenswrapper[4744]: E1008 09:47:41.601243 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae54bee9e483720e48891c1ab3de14273dbfe975609e075770b8976e6da89136\": container with ID starting with ae54bee9e483720e48891c1ab3de14273dbfe975609e075770b8976e6da89136 not found: ID does not exist" containerID="ae54bee9e483720e48891c1ab3de14273dbfe975609e075770b8976e6da89136" Oct 08 09:47:41 crc kubenswrapper[4744]: I1008 09:47:41.601417 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae54bee9e483720e48891c1ab3de14273dbfe975609e075770b8976e6da89136"} err="failed to get container status \"ae54bee9e483720e48891c1ab3de14273dbfe975609e075770b8976e6da89136\": rpc error: code = NotFound desc = could not find container \"ae54bee9e483720e48891c1ab3de14273dbfe975609e075770b8976e6da89136\": container with ID starting with ae54bee9e483720e48891c1ab3de14273dbfe975609e075770b8976e6da89136 not found: ID does not exist" Oct 08 09:47:43 crc kubenswrapper[4744]: I1008 09:47:43.464895 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="154a234a-1453-4689-9afe-00c47172bd87" path="/var/lib/kubelet/pods/154a234a-1453-4689-9afe-00c47172bd87/volumes" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.266679 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96"] Oct 08 09:47:44 crc kubenswrapper[4744]: E1008 09:47:44.267156 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154a234a-1453-4689-9afe-00c47172bd87" containerName="extract-utilities" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.267176 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="154a234a-1453-4689-9afe-00c47172bd87" containerName="extract-utilities" Oct 08 09:47:44 crc kubenswrapper[4744]: E1008 09:47:44.267218 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154a234a-1453-4689-9afe-00c47172bd87" containerName="extract-content" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.267228 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="154a234a-1453-4689-9afe-00c47172bd87" containerName="extract-content" Oct 08 09:47:44 crc kubenswrapper[4744]: E1008 09:47:44.267259 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154a234a-1453-4689-9afe-00c47172bd87" containerName="registry-server" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.267269 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="154a234a-1453-4689-9afe-00c47172bd87" containerName="registry-server" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.267505 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="154a234a-1453-4689-9afe-00c47172bd87" containerName="registry-server" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.268322 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.271477 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.273037 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.273165 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.273451 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.276978 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.326697 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96"] Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.344181 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.344235 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.344273 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.344323 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75qzn\" (UniqueName: \"kubernetes.io/projected/50611401-ad48-4b64-ad34-48a7f2d71a11-kube-api-access-75qzn\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.344403 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.446589 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.446690 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.446723 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.446755 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.446806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75qzn\" (UniqueName: \"kubernetes.io/projected/50611401-ad48-4b64-ad34-48a7f2d71a11-kube-api-access-75qzn\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.454546 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.455300 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.456195 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.457398 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.470277 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75qzn\" (UniqueName: \"kubernetes.io/projected/50611401-ad48-4b64-ad34-48a7f2d71a11-kube-api-access-75qzn\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:44 crc kubenswrapper[4744]: I1008 09:47:44.590672 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:47:45 crc kubenswrapper[4744]: I1008 09:47:45.172510 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96"] Oct 08 09:47:45 crc kubenswrapper[4744]: I1008 09:47:45.517334 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" event={"ID":"50611401-ad48-4b64-ad34-48a7f2d71a11","Type":"ContainerStarted","Data":"658d6768eb3ddce8f6ff3292e58056dc711b907f31c03a7da33303a95150f625"} Oct 08 09:47:45 crc kubenswrapper[4744]: I1008 09:47:45.698820 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:47:46 crc kubenswrapper[4744]: I1008 09:47:46.533019 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" event={"ID":"50611401-ad48-4b64-ad34-48a7f2d71a11","Type":"ContainerStarted","Data":"7091c070170c9941229027d4722ee6872c24c40e498555b6421082c579fd8a70"} Oct 08 09:47:46 crc kubenswrapper[4744]: I1008 09:47:46.556294 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" podStartSLOduration=2.040869143 podStartE2EDuration="2.556273902s" podCreationTimestamp="2025-10-08 09:47:44 +0000 UTC" firstStartedPulling="2025-10-08 09:47:45.179750058 +0000 UTC m=+2160.427395307" lastFinishedPulling="2025-10-08 09:47:45.695154787 +0000 UTC m=+2160.942800066" observedRunningTime="2025-10-08 09:47:46.552494626 +0000 UTC m=+2161.800139865" watchObservedRunningTime="2025-10-08 09:47:46.556273902 +0000 UTC m=+2161.803919141" Oct 08 09:47:55 crc kubenswrapper[4744]: I1008 09:47:55.357009 4744 scope.go:117] "RemoveContainer" containerID="2c1bc45b415c831d02f110d8022cf5178bc6d2438fb903f0b5eef8497be69431" Oct 08 09:47:55 crc kubenswrapper[4744]: I1008 09:47:55.410575 4744 scope.go:117] "RemoveContainer" containerID="89a1061ef4c1b296eaa022b9c64167e6025c997579d6349ca258a18cac3d0b7f" Oct 08 09:47:55 crc kubenswrapper[4744]: I1008 09:47:55.450290 4744 scope.go:117] "RemoveContainer" containerID="a875f357e53d04f7ad56db08a071dc57998c61748d053575531c0fa580986d56" Oct 08 09:47:55 crc kubenswrapper[4744]: I1008 09:47:55.493432 4744 scope.go:117] "RemoveContainer" containerID="8aa46e31f25cdc651f531c404f2421a2c12c7db27284eb48602697e74262cf42" Oct 08 09:47:55 crc kubenswrapper[4744]: I1008 09:47:55.562456 4744 scope.go:117] "RemoveContainer" containerID="97478bbac68c97a342195072be4524df21c40d0ff001858bfc07b16e281d4b9a" Oct 08 09:47:55 crc kubenswrapper[4744]: I1008 09:47:55.622571 4744 scope.go:117] "RemoveContainer" containerID="2ee85605b95bfb7b4c80f58f3e7d93f8c30f3a39fb7bbdd463384c8c50eea904" Oct 08 09:47:55 crc kubenswrapper[4744]: I1008 09:47:55.690607 4744 scope.go:117] "RemoveContainer" containerID="7c29bfe4edffea256e1fdc8047cb22421a81ce926b13da96a82073a4dcd53a25" Oct 08 09:47:55 crc kubenswrapper[4744]: I1008 09:47:55.735719 4744 scope.go:117] "RemoveContainer" containerID="34274e263382b096e9cec1e24525e4ef27c4003bad261a860a35e16df357fd63" Oct 08 09:47:55 crc kubenswrapper[4744]: I1008 09:47:55.804682 4744 scope.go:117] "RemoveContainer" containerID="9e07c94b16ead85a8580ea2841cf7884b57509e2d083a5ab1bb80ff46a16c36a" Oct 08 09:47:55 crc kubenswrapper[4744]: I1008 09:47:55.845069 4744 scope.go:117] "RemoveContainer" containerID="e87c8e3bb3965b78b6ec142165e56ed4fe5c61c052fd8f46119a17d9b77346d9" Oct 08 09:47:55 crc kubenswrapper[4744]: I1008 09:47:55.883187 4744 scope.go:117] "RemoveContainer" containerID="1b67b47da4a52585a88d6873c6a6f7a8bb324cb24c8579f5bb9a0fc23cbaf42e" Oct 08 09:47:58 crc kubenswrapper[4744]: I1008 09:47:58.688465 4744 generic.go:334] "Generic (PLEG): container finished" podID="50611401-ad48-4b64-ad34-48a7f2d71a11" containerID="7091c070170c9941229027d4722ee6872c24c40e498555b6421082c579fd8a70" exitCode=0 Oct 08 09:47:58 crc kubenswrapper[4744]: I1008 09:47:58.688528 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" event={"ID":"50611401-ad48-4b64-ad34-48a7f2d71a11","Type":"ContainerDied","Data":"7091c070170c9941229027d4722ee6872c24c40e498555b6421082c579fd8a70"} Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.153535 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.306591 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-repo-setup-combined-ca-bundle\") pod \"50611401-ad48-4b64-ad34-48a7f2d71a11\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.306772 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-ceph\") pod \"50611401-ad48-4b64-ad34-48a7f2d71a11\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.306821 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-ssh-key\") pod \"50611401-ad48-4b64-ad34-48a7f2d71a11\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.306866 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75qzn\" (UniqueName: \"kubernetes.io/projected/50611401-ad48-4b64-ad34-48a7f2d71a11-kube-api-access-75qzn\") pod \"50611401-ad48-4b64-ad34-48a7f2d71a11\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.306985 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-inventory\") pod \"50611401-ad48-4b64-ad34-48a7f2d71a11\" (UID: \"50611401-ad48-4b64-ad34-48a7f2d71a11\") " Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.313094 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50611401-ad48-4b64-ad34-48a7f2d71a11-kube-api-access-75qzn" (OuterVolumeSpecName: "kube-api-access-75qzn") pod "50611401-ad48-4b64-ad34-48a7f2d71a11" (UID: "50611401-ad48-4b64-ad34-48a7f2d71a11"). InnerVolumeSpecName "kube-api-access-75qzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.320214 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-ceph" (OuterVolumeSpecName: "ceph") pod "50611401-ad48-4b64-ad34-48a7f2d71a11" (UID: "50611401-ad48-4b64-ad34-48a7f2d71a11"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.320315 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "50611401-ad48-4b64-ad34-48a7f2d71a11" (UID: "50611401-ad48-4b64-ad34-48a7f2d71a11"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.342048 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-inventory" (OuterVolumeSpecName: "inventory") pod "50611401-ad48-4b64-ad34-48a7f2d71a11" (UID: "50611401-ad48-4b64-ad34-48a7f2d71a11"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.342141 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "50611401-ad48-4b64-ad34-48a7f2d71a11" (UID: "50611401-ad48-4b64-ad34-48a7f2d71a11"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.408763 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.408806 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75qzn\" (UniqueName: \"kubernetes.io/projected/50611401-ad48-4b64-ad34-48a7f2d71a11-kube-api-access-75qzn\") on node \"crc\" DevicePath \"\"" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.408817 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.408826 4744 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.408834 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50611401-ad48-4b64-ad34-48a7f2d71a11-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.728997 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" event={"ID":"50611401-ad48-4b64-ad34-48a7f2d71a11","Type":"ContainerDied","Data":"658d6768eb3ddce8f6ff3292e58056dc711b907f31c03a7da33303a95150f625"} Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.729069 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="658d6768eb3ddce8f6ff3292e58056dc711b907f31c03a7da33303a95150f625" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.729119 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.811082 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4"] Oct 08 09:48:00 crc kubenswrapper[4744]: E1008 09:48:00.811491 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50611401-ad48-4b64-ad34-48a7f2d71a11" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.811506 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="50611401-ad48-4b64-ad34-48a7f2d71a11" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.811693 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="50611401-ad48-4b64-ad34-48a7f2d71a11" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.812347 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.814884 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.815452 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.815717 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.815814 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.830745 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4"] Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.833472 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.949330 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.949452 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.949481 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcsqn\" (UniqueName: \"kubernetes.io/projected/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-kube-api-access-hcsqn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.949515 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:00 crc kubenswrapper[4744]: I1008 09:48:00.949786 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.051237 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.051327 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcsqn\" (UniqueName: \"kubernetes.io/projected/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-kube-api-access-hcsqn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.051350 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.051399 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.051453 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.057479 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.058334 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.059487 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.061534 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.076480 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcsqn\" (UniqueName: \"kubernetes.io/projected/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-kube-api-access-hcsqn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.133588 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.690547 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4"] Oct 08 09:48:01 crc kubenswrapper[4744]: I1008 09:48:01.742637 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" event={"ID":"f0464e55-2b52-4c5a-bc87-9c2c87cb6101","Type":"ContainerStarted","Data":"05e094681f4d6f6baada998cc44384f544403b0487db3e52bb0459efe83a32b5"} Oct 08 09:48:02 crc kubenswrapper[4744]: I1008 09:48:02.752435 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" event={"ID":"f0464e55-2b52-4c5a-bc87-9c2c87cb6101","Type":"ContainerStarted","Data":"5e42aeb351d3a7169359ff9207b6340bda459f72702e0fa1c8a709b3541d0c4f"} Oct 08 09:48:02 crc kubenswrapper[4744]: I1008 09:48:02.774845 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" podStartSLOduration=2.258039002 podStartE2EDuration="2.774819599s" podCreationTimestamp="2025-10-08 09:48:00 +0000 UTC" firstStartedPulling="2025-10-08 09:48:01.70867935 +0000 UTC m=+2176.956324589" lastFinishedPulling="2025-10-08 09:48:02.225459947 +0000 UTC m=+2177.473105186" observedRunningTime="2025-10-08 09:48:02.770979093 +0000 UTC m=+2178.018624332" watchObservedRunningTime="2025-10-08 09:48:02.774819599 +0000 UTC m=+2178.022464858" Oct 08 09:48:49 crc kubenswrapper[4744]: I1008 09:48:49.690819 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:48:49 crc kubenswrapper[4744]: I1008 09:48:49.691823 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:49:19 crc kubenswrapper[4744]: I1008 09:49:19.690665 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:49:19 crc kubenswrapper[4744]: I1008 09:49:19.691203 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:49:44 crc kubenswrapper[4744]: I1008 09:49:44.623974 4744 generic.go:334] "Generic (PLEG): container finished" podID="f0464e55-2b52-4c5a-bc87-9c2c87cb6101" containerID="5e42aeb351d3a7169359ff9207b6340bda459f72702e0fa1c8a709b3541d0c4f" exitCode=0 Oct 08 09:49:44 crc kubenswrapper[4744]: I1008 09:49:44.624026 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" event={"ID":"f0464e55-2b52-4c5a-bc87-9c2c87cb6101","Type":"ContainerDied","Data":"5e42aeb351d3a7169359ff9207b6340bda459f72702e0fa1c8a709b3541d0c4f"} Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.071840 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.240822 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-ssh-key\") pod \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.240979 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-bootstrap-combined-ca-bundle\") pod \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.241071 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-inventory\") pod \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.241164 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcsqn\" (UniqueName: \"kubernetes.io/projected/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-kube-api-access-hcsqn\") pod \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.241207 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-ceph\") pod \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\" (UID: \"f0464e55-2b52-4c5a-bc87-9c2c87cb6101\") " Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.264829 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-kube-api-access-hcsqn" (OuterVolumeSpecName: "kube-api-access-hcsqn") pod "f0464e55-2b52-4c5a-bc87-9c2c87cb6101" (UID: "f0464e55-2b52-4c5a-bc87-9c2c87cb6101"). InnerVolumeSpecName "kube-api-access-hcsqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.269641 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f0464e55-2b52-4c5a-bc87-9c2c87cb6101" (UID: "f0464e55-2b52-4c5a-bc87-9c2c87cb6101"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.272889 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-inventory" (OuterVolumeSpecName: "inventory") pod "f0464e55-2b52-4c5a-bc87-9c2c87cb6101" (UID: "f0464e55-2b52-4c5a-bc87-9c2c87cb6101"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.275436 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-ceph" (OuterVolumeSpecName: "ceph") pod "f0464e55-2b52-4c5a-bc87-9c2c87cb6101" (UID: "f0464e55-2b52-4c5a-bc87-9c2c87cb6101"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.287489 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f0464e55-2b52-4c5a-bc87-9c2c87cb6101" (UID: "f0464e55-2b52-4c5a-bc87-9c2c87cb6101"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.343803 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.343852 4744 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.343873 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.343888 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcsqn\" (UniqueName: \"kubernetes.io/projected/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-kube-api-access-hcsqn\") on node \"crc\" DevicePath \"\"" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.343903 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f0464e55-2b52-4c5a-bc87-9c2c87cb6101-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.641517 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" event={"ID":"f0464e55-2b52-4c5a-bc87-9c2c87cb6101","Type":"ContainerDied","Data":"05e094681f4d6f6baada998cc44384f544403b0487db3e52bb0459efe83a32b5"} Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.641563 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05e094681f4d6f6baada998cc44384f544403b0487db3e52bb0459efe83a32b5" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.641570 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.759558 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv"] Oct 08 09:49:46 crc kubenswrapper[4744]: E1008 09:49:46.760022 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0464e55-2b52-4c5a-bc87-9c2c87cb6101" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.760044 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0464e55-2b52-4c5a-bc87-9c2c87cb6101" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.760307 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0464e55-2b52-4c5a-bc87-9c2c87cb6101" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.761257 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.768786 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.768880 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.768892 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.769050 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.771545 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.786687 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv"] Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.851227 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6l7x\" (UniqueName: \"kubernetes.io/projected/d0fc8259-e8cd-4d69-8a95-f74d908d569e-kube-api-access-v6l7x\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k48qv\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.851556 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k48qv\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.851578 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k48qv\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.851619 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k48qv\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.952997 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k48qv\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.953078 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6l7x\" (UniqueName: \"kubernetes.io/projected/d0fc8259-e8cd-4d69-8a95-f74d908d569e-kube-api-access-v6l7x\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k48qv\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.953168 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k48qv\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.953188 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k48qv\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.957208 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k48qv\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.957247 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k48qv\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.957644 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k48qv\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:46 crc kubenswrapper[4744]: I1008 09:49:46.975086 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6l7x\" (UniqueName: \"kubernetes.io/projected/d0fc8259-e8cd-4d69-8a95-f74d908d569e-kube-api-access-v6l7x\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-k48qv\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:47 crc kubenswrapper[4744]: I1008 09:49:47.085452 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:49:47 crc kubenswrapper[4744]: I1008 09:49:47.640190 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv"] Oct 08 09:49:47 crc kubenswrapper[4744]: I1008 09:49:47.652125 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 09:49:48 crc kubenswrapper[4744]: I1008 09:49:48.661629 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" event={"ID":"d0fc8259-e8cd-4d69-8a95-f74d908d569e","Type":"ContainerStarted","Data":"b1eb423dbf0a5bf3b68b977a016c3ab4cf5a9ea777cca99496de1a9018a3c8ba"} Oct 08 09:49:48 crc kubenswrapper[4744]: I1008 09:49:48.661959 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" event={"ID":"d0fc8259-e8cd-4d69-8a95-f74d908d569e","Type":"ContainerStarted","Data":"ea75806f0fd54403aabd16beaaf3c4b35e2dbc590cb421520d0a0a9e9c08b5c2"} Oct 08 09:49:48 crc kubenswrapper[4744]: I1008 09:49:48.685691 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" podStartSLOduration=2.23511579 podStartE2EDuration="2.685668012s" podCreationTimestamp="2025-10-08 09:49:46 +0000 UTC" firstStartedPulling="2025-10-08 09:49:47.651920088 +0000 UTC m=+2282.899565327" lastFinishedPulling="2025-10-08 09:49:48.10247231 +0000 UTC m=+2283.350117549" observedRunningTime="2025-10-08 09:49:48.680777315 +0000 UTC m=+2283.928422554" watchObservedRunningTime="2025-10-08 09:49:48.685668012 +0000 UTC m=+2283.933313251" Oct 08 09:49:49 crc kubenswrapper[4744]: I1008 09:49:49.690962 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:49:49 crc kubenswrapper[4744]: I1008 09:49:49.691301 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:49:49 crc kubenswrapper[4744]: I1008 09:49:49.691356 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:49:49 crc kubenswrapper[4744]: I1008 09:49:49.692158 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 09:49:49 crc kubenswrapper[4744]: I1008 09:49:49.692219 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" gracePeriod=600 Oct 08 09:49:49 crc kubenswrapper[4744]: E1008 09:49:49.841403 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:49:50 crc kubenswrapper[4744]: I1008 09:49:50.680763 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" exitCode=0 Oct 08 09:49:50 crc kubenswrapper[4744]: I1008 09:49:50.680878 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af"} Oct 08 09:49:50 crc kubenswrapper[4744]: I1008 09:49:50.681322 4744 scope.go:117] "RemoveContainer" containerID="edfca7ab5f128667cb0425a5ce4b92f849206f6e3061358fe9d17c2c761c3416" Oct 08 09:49:50 crc kubenswrapper[4744]: I1008 09:49:50.681996 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:49:50 crc kubenswrapper[4744]: E1008 09:49:50.682279 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:50:05 crc kubenswrapper[4744]: I1008 09:50:05.457520 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:50:05 crc kubenswrapper[4744]: E1008 09:50:05.459899 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:50:14 crc kubenswrapper[4744]: I1008 09:50:14.934978 4744 generic.go:334] "Generic (PLEG): container finished" podID="d0fc8259-e8cd-4d69-8a95-f74d908d569e" containerID="b1eb423dbf0a5bf3b68b977a016c3ab4cf5a9ea777cca99496de1a9018a3c8ba" exitCode=0 Oct 08 09:50:14 crc kubenswrapper[4744]: I1008 09:50:14.935066 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" event={"ID":"d0fc8259-e8cd-4d69-8a95-f74d908d569e","Type":"ContainerDied","Data":"b1eb423dbf0a5bf3b68b977a016c3ab4cf5a9ea777cca99496de1a9018a3c8ba"} Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.362568 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.456010 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-ssh-key\") pod \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.456049 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-inventory\") pod \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.456119 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-ceph\") pod \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.456158 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6l7x\" (UniqueName: \"kubernetes.io/projected/d0fc8259-e8cd-4d69-8a95-f74d908d569e-kube-api-access-v6l7x\") pod \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\" (UID: \"d0fc8259-e8cd-4d69-8a95-f74d908d569e\") " Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.463738 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-ceph" (OuterVolumeSpecName: "ceph") pod "d0fc8259-e8cd-4d69-8a95-f74d908d569e" (UID: "d0fc8259-e8cd-4d69-8a95-f74d908d569e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.465235 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0fc8259-e8cd-4d69-8a95-f74d908d569e-kube-api-access-v6l7x" (OuterVolumeSpecName: "kube-api-access-v6l7x") pod "d0fc8259-e8cd-4d69-8a95-f74d908d569e" (UID: "d0fc8259-e8cd-4d69-8a95-f74d908d569e"). InnerVolumeSpecName "kube-api-access-v6l7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.483737 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d0fc8259-e8cd-4d69-8a95-f74d908d569e" (UID: "d0fc8259-e8cd-4d69-8a95-f74d908d569e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.491253 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-inventory" (OuterVolumeSpecName: "inventory") pod "d0fc8259-e8cd-4d69-8a95-f74d908d569e" (UID: "d0fc8259-e8cd-4d69-8a95-f74d908d569e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.558028 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.558055 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.558064 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d0fc8259-e8cd-4d69-8a95-f74d908d569e-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.558076 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6l7x\" (UniqueName: \"kubernetes.io/projected/d0fc8259-e8cd-4d69-8a95-f74d908d569e-kube-api-access-v6l7x\") on node \"crc\" DevicePath \"\"" Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.954054 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" event={"ID":"d0fc8259-e8cd-4d69-8a95-f74d908d569e","Type":"ContainerDied","Data":"ea75806f0fd54403aabd16beaaf3c4b35e2dbc590cb421520d0a0a9e9c08b5c2"} Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.954096 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea75806f0fd54403aabd16beaaf3c4b35e2dbc590cb421520d0a0a9e9c08b5c2" Oct 08 09:50:16 crc kubenswrapper[4744]: I1008 09:50:16.954147 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-k48qv" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.055627 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh"] Oct 08 09:50:17 crc kubenswrapper[4744]: E1008 09:50:17.056047 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0fc8259-e8cd-4d69-8a95-f74d908d569e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.056068 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0fc8259-e8cd-4d69-8a95-f74d908d569e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.056296 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0fc8259-e8cd-4d69-8a95-f74d908d569e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.057105 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.062976 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.063153 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh"] Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.064814 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.066265 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.066605 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.068250 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.167767 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfv48\" (UniqueName: \"kubernetes.io/projected/4e3fa5bc-5c8b-45de-a173-daf069f2868f-kube-api-access-hfv48\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.167847 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.168207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.168276 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.270626 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfv48\" (UniqueName: \"kubernetes.io/projected/4e3fa5bc-5c8b-45de-a173-daf069f2868f-kube-api-access-hfv48\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.270705 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.270777 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.270796 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.276308 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.278283 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.282360 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.302187 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfv48\" (UniqueName: \"kubernetes.io/projected/4e3fa5bc-5c8b-45de-a173-daf069f2868f-kube-api-access-hfv48\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:17 crc kubenswrapper[4744]: I1008 09:50:17.411716 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:18 crc kubenswrapper[4744]: I1008 09:50:18.013319 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh"] Oct 08 09:50:18 crc kubenswrapper[4744]: I1008 09:50:18.971349 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" event={"ID":"4e3fa5bc-5c8b-45de-a173-daf069f2868f","Type":"ContainerStarted","Data":"245c313dd9cf40b5cd708d0b3fbc895e20fc2fa69fd5450dcfa17339ea19a448"} Oct 08 09:50:18 crc kubenswrapper[4744]: I1008 09:50:18.971674 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" event={"ID":"4e3fa5bc-5c8b-45de-a173-daf069f2868f","Type":"ContainerStarted","Data":"cec4c7f1f5d55cfd68a186c30bdc509b36ee6c49fe715c58c57c70c48fb9fc10"} Oct 08 09:50:19 crc kubenswrapper[4744]: I1008 09:50:19.453030 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:50:19 crc kubenswrapper[4744]: E1008 09:50:19.453342 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:50:25 crc kubenswrapper[4744]: I1008 09:50:25.033791 4744 generic.go:334] "Generic (PLEG): container finished" podID="4e3fa5bc-5c8b-45de-a173-daf069f2868f" containerID="245c313dd9cf40b5cd708d0b3fbc895e20fc2fa69fd5450dcfa17339ea19a448" exitCode=0 Oct 08 09:50:25 crc kubenswrapper[4744]: I1008 09:50:25.033828 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" event={"ID":"4e3fa5bc-5c8b-45de-a173-daf069f2868f","Type":"ContainerDied","Data":"245c313dd9cf40b5cd708d0b3fbc895e20fc2fa69fd5450dcfa17339ea19a448"} Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.441078 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.484821 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-inventory\") pod \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.485189 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfv48\" (UniqueName: \"kubernetes.io/projected/4e3fa5bc-5c8b-45de-a173-daf069f2868f-kube-api-access-hfv48\") pod \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.485311 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-ceph\") pod \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.485349 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-ssh-key\") pod \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\" (UID: \"4e3fa5bc-5c8b-45de-a173-daf069f2868f\") " Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.494064 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-ceph" (OuterVolumeSpecName: "ceph") pod "4e3fa5bc-5c8b-45de-a173-daf069f2868f" (UID: "4e3fa5bc-5c8b-45de-a173-daf069f2868f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.506801 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e3fa5bc-5c8b-45de-a173-daf069f2868f-kube-api-access-hfv48" (OuterVolumeSpecName: "kube-api-access-hfv48") pod "4e3fa5bc-5c8b-45de-a173-daf069f2868f" (UID: "4e3fa5bc-5c8b-45de-a173-daf069f2868f"). InnerVolumeSpecName "kube-api-access-hfv48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.538523 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-inventory" (OuterVolumeSpecName: "inventory") pod "4e3fa5bc-5c8b-45de-a173-daf069f2868f" (UID: "4e3fa5bc-5c8b-45de-a173-daf069f2868f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.541569 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4e3fa5bc-5c8b-45de-a173-daf069f2868f" (UID: "4e3fa5bc-5c8b-45de-a173-daf069f2868f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.587837 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.587876 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfv48\" (UniqueName: \"kubernetes.io/projected/4e3fa5bc-5c8b-45de-a173-daf069f2868f-kube-api-access-hfv48\") on node \"crc\" DevicePath \"\"" Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.587891 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:50:26 crc kubenswrapper[4744]: I1008 09:50:26.587901 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e3fa5bc-5c8b-45de-a173-daf069f2868f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.052503 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" event={"ID":"4e3fa5bc-5c8b-45de-a173-daf069f2868f","Type":"ContainerDied","Data":"cec4c7f1f5d55cfd68a186c30bdc509b36ee6c49fe715c58c57c70c48fb9fc10"} Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.052768 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cec4c7f1f5d55cfd68a186c30bdc509b36ee6c49fe715c58c57c70c48fb9fc10" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.052877 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.141914 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks"] Oct 08 09:50:27 crc kubenswrapper[4744]: E1008 09:50:27.142300 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e3fa5bc-5c8b-45de-a173-daf069f2868f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.142319 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e3fa5bc-5c8b-45de-a173-daf069f2868f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.142504 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e3fa5bc-5c8b-45de-a173-daf069f2868f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.143225 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.145534 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.145684 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.145691 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.145828 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.146065 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.148052 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks"] Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.303836 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cd6ks\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.304011 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cd6ks\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.304044 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cd6ks\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.304117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zvmf\" (UniqueName: \"kubernetes.io/projected/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-kube-api-access-4zvmf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cd6ks\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.405301 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cd6ks\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.405350 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cd6ks\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.405420 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zvmf\" (UniqueName: \"kubernetes.io/projected/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-kube-api-access-4zvmf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cd6ks\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.405484 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cd6ks\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.410736 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cd6ks\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.411060 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cd6ks\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.418845 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cd6ks\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.429793 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zvmf\" (UniqueName: \"kubernetes.io/projected/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-kube-api-access-4zvmf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cd6ks\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:27 crc kubenswrapper[4744]: I1008 09:50:27.507893 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:50:28 crc kubenswrapper[4744]: I1008 09:50:28.063042 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks"] Oct 08 09:50:29 crc kubenswrapper[4744]: I1008 09:50:29.067755 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" event={"ID":"57e3beb8-aeaa-435a-aa5a-78f854db1d6d","Type":"ContainerStarted","Data":"facb4aa6eeb8f31ff26b6a4b649b6ec462c1ec77e7fad459c301cedbc3a77c27"} Oct 08 09:50:29 crc kubenswrapper[4744]: I1008 09:50:29.068099 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" event={"ID":"57e3beb8-aeaa-435a-aa5a-78f854db1d6d","Type":"ContainerStarted","Data":"ebcb3b5d5a6355df8a15974835423cdb5e72edb59c4a88225ecd5d3e39ab19c2"} Oct 08 09:50:29 crc kubenswrapper[4744]: I1008 09:50:29.081732 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" podStartSLOduration=1.483760402 podStartE2EDuration="2.081713244s" podCreationTimestamp="2025-10-08 09:50:27 +0000 UTC" firstStartedPulling="2025-10-08 09:50:28.06776694 +0000 UTC m=+2323.315412179" lastFinishedPulling="2025-10-08 09:50:28.665719782 +0000 UTC m=+2323.913365021" observedRunningTime="2025-10-08 09:50:29.078981068 +0000 UTC m=+2324.326626297" watchObservedRunningTime="2025-10-08 09:50:29.081713244 +0000 UTC m=+2324.329358483" Oct 08 09:50:30 crc kubenswrapper[4744]: I1008 09:50:30.453484 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:50:30 crc kubenswrapper[4744]: E1008 09:50:30.454092 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:50:45 crc kubenswrapper[4744]: I1008 09:50:45.457397 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:50:45 crc kubenswrapper[4744]: E1008 09:50:45.458218 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:51:00 crc kubenswrapper[4744]: I1008 09:51:00.452873 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:51:00 crc kubenswrapper[4744]: E1008 09:51:00.453558 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:51:06 crc kubenswrapper[4744]: I1008 09:51:06.359669 4744 generic.go:334] "Generic (PLEG): container finished" podID="57e3beb8-aeaa-435a-aa5a-78f854db1d6d" containerID="facb4aa6eeb8f31ff26b6a4b649b6ec462c1ec77e7fad459c301cedbc3a77c27" exitCode=0 Oct 08 09:51:06 crc kubenswrapper[4744]: I1008 09:51:06.359788 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" event={"ID":"57e3beb8-aeaa-435a-aa5a-78f854db1d6d","Type":"ContainerDied","Data":"facb4aa6eeb8f31ff26b6a4b649b6ec462c1ec77e7fad459c301cedbc3a77c27"} Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.735428 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.787328 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-ceph\") pod \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.787364 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-ssh-key\") pod \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.787393 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-inventory\") pod \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.787412 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zvmf\" (UniqueName: \"kubernetes.io/projected/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-kube-api-access-4zvmf\") pod \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\" (UID: \"57e3beb8-aeaa-435a-aa5a-78f854db1d6d\") " Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.792717 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-kube-api-access-4zvmf" (OuterVolumeSpecName: "kube-api-access-4zvmf") pod "57e3beb8-aeaa-435a-aa5a-78f854db1d6d" (UID: "57e3beb8-aeaa-435a-aa5a-78f854db1d6d"). InnerVolumeSpecName "kube-api-access-4zvmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.793458 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-ceph" (OuterVolumeSpecName: "ceph") pod "57e3beb8-aeaa-435a-aa5a-78f854db1d6d" (UID: "57e3beb8-aeaa-435a-aa5a-78f854db1d6d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.812776 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-inventory" (OuterVolumeSpecName: "inventory") pod "57e3beb8-aeaa-435a-aa5a-78f854db1d6d" (UID: "57e3beb8-aeaa-435a-aa5a-78f854db1d6d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.815586 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "57e3beb8-aeaa-435a-aa5a-78f854db1d6d" (UID: "57e3beb8-aeaa-435a-aa5a-78f854db1d6d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.888884 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.888916 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.888927 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:51:07 crc kubenswrapper[4744]: I1008 09:51:07.888936 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zvmf\" (UniqueName: \"kubernetes.io/projected/57e3beb8-aeaa-435a-aa5a-78f854db1d6d-kube-api-access-4zvmf\") on node \"crc\" DevicePath \"\"" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.377619 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" event={"ID":"57e3beb8-aeaa-435a-aa5a-78f854db1d6d","Type":"ContainerDied","Data":"ebcb3b5d5a6355df8a15974835423cdb5e72edb59c4a88225ecd5d3e39ab19c2"} Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.378095 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebcb3b5d5a6355df8a15974835423cdb5e72edb59c4a88225ecd5d3e39ab19c2" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.377659 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cd6ks" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.480940 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl"] Oct 08 09:51:08 crc kubenswrapper[4744]: E1008 09:51:08.481642 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e3beb8-aeaa-435a-aa5a-78f854db1d6d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.481673 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e3beb8-aeaa-435a-aa5a-78f854db1d6d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.481952 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="57e3beb8-aeaa-435a-aa5a-78f854db1d6d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.482862 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.488123 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl"] Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.508177 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.508706 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.509643 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.509715 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.509964 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.613339 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqmqm\" (UniqueName: \"kubernetes.io/projected/dd384bed-7c22-4e1a-8342-94589991c934-kube-api-access-vqmqm\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.613433 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.613479 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.613610 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.714986 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.715117 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqmqm\" (UniqueName: \"kubernetes.io/projected/dd384bed-7c22-4e1a-8342-94589991c934-kube-api-access-vqmqm\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.715172 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.715210 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.720232 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.721554 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.723466 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.742037 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqmqm\" (UniqueName: \"kubernetes.io/projected/dd384bed-7c22-4e1a-8342-94589991c934-kube-api-access-vqmqm\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:08 crc kubenswrapper[4744]: I1008 09:51:08.832257 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:09 crc kubenswrapper[4744]: I1008 09:51:09.393784 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl"] Oct 08 09:51:10 crc kubenswrapper[4744]: I1008 09:51:10.400483 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" event={"ID":"dd384bed-7c22-4e1a-8342-94589991c934","Type":"ContainerStarted","Data":"acffbf188d368306be3d9ea5d93dc05e6610dd15f641963050e318fdf5738e6f"} Oct 08 09:51:10 crc kubenswrapper[4744]: I1008 09:51:10.400998 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" event={"ID":"dd384bed-7c22-4e1a-8342-94589991c934","Type":"ContainerStarted","Data":"88daf93bf84b202a79aba7fe95635f5c25258f2b59407021613a0168f6ea90cf"} Oct 08 09:51:10 crc kubenswrapper[4744]: I1008 09:51:10.424898 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" podStartSLOduration=1.902211683 podStartE2EDuration="2.42488072s" podCreationTimestamp="2025-10-08 09:51:08 +0000 UTC" firstStartedPulling="2025-10-08 09:51:09.407208794 +0000 UTC m=+2364.654854053" lastFinishedPulling="2025-10-08 09:51:09.929877841 +0000 UTC m=+2365.177523090" observedRunningTime="2025-10-08 09:51:10.420254931 +0000 UTC m=+2365.667900180" watchObservedRunningTime="2025-10-08 09:51:10.42488072 +0000 UTC m=+2365.672525959" Oct 08 09:51:14 crc kubenswrapper[4744]: I1008 09:51:14.439598 4744 generic.go:334] "Generic (PLEG): container finished" podID="dd384bed-7c22-4e1a-8342-94589991c934" containerID="acffbf188d368306be3d9ea5d93dc05e6610dd15f641963050e318fdf5738e6f" exitCode=0 Oct 08 09:51:14 crc kubenswrapper[4744]: I1008 09:51:14.439679 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" event={"ID":"dd384bed-7c22-4e1a-8342-94589991c934","Type":"ContainerDied","Data":"acffbf188d368306be3d9ea5d93dc05e6610dd15f641963050e318fdf5738e6f"} Oct 08 09:51:15 crc kubenswrapper[4744]: I1008 09:51:15.458048 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:51:15 crc kubenswrapper[4744]: E1008 09:51:15.459024 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.000662 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.058468 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-ssh-key\") pod \"dd384bed-7c22-4e1a-8342-94589991c934\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.058608 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqmqm\" (UniqueName: \"kubernetes.io/projected/dd384bed-7c22-4e1a-8342-94589991c934-kube-api-access-vqmqm\") pod \"dd384bed-7c22-4e1a-8342-94589991c934\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.058735 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-ceph\") pod \"dd384bed-7c22-4e1a-8342-94589991c934\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.058797 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-inventory\") pod \"dd384bed-7c22-4e1a-8342-94589991c934\" (UID: \"dd384bed-7c22-4e1a-8342-94589991c934\") " Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.064765 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd384bed-7c22-4e1a-8342-94589991c934-kube-api-access-vqmqm" (OuterVolumeSpecName: "kube-api-access-vqmqm") pod "dd384bed-7c22-4e1a-8342-94589991c934" (UID: "dd384bed-7c22-4e1a-8342-94589991c934"). InnerVolumeSpecName "kube-api-access-vqmqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.085756 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dd384bed-7c22-4e1a-8342-94589991c934" (UID: "dd384bed-7c22-4e1a-8342-94589991c934"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.089306 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-inventory" (OuterVolumeSpecName: "inventory") pod "dd384bed-7c22-4e1a-8342-94589991c934" (UID: "dd384bed-7c22-4e1a-8342-94589991c934"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.091708 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-ceph" (OuterVolumeSpecName: "ceph") pod "dd384bed-7c22-4e1a-8342-94589991c934" (UID: "dd384bed-7c22-4e1a-8342-94589991c934"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.161527 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.161589 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.161606 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqmqm\" (UniqueName: \"kubernetes.io/projected/dd384bed-7c22-4e1a-8342-94589991c934-kube-api-access-vqmqm\") on node \"crc\" DevicePath \"\"" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.161631 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd384bed-7c22-4e1a-8342-94589991c934-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.460034 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" event={"ID":"dd384bed-7c22-4e1a-8342-94589991c934","Type":"ContainerDied","Data":"88daf93bf84b202a79aba7fe95635f5c25258f2b59407021613a0168f6ea90cf"} Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.460070 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88daf93bf84b202a79aba7fe95635f5c25258f2b59407021613a0168f6ea90cf" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.460104 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.554264 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57"] Oct 08 09:51:16 crc kubenswrapper[4744]: E1008 09:51:16.554927 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd384bed-7c22-4e1a-8342-94589991c934" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.554979 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd384bed-7c22-4e1a-8342-94589991c934" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.555564 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd384bed-7c22-4e1a-8342-94589991c934" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.556515 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.558360 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.562616 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.562855 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.562887 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.563179 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.587610 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57"] Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.669422 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wcf57\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.669486 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvgvx\" (UniqueName: \"kubernetes.io/projected/bf19b11f-c43e-4f3b-a34c-53d3837705fb-kube-api-access-qvgvx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wcf57\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.669554 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wcf57\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.669594 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wcf57\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.770796 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wcf57\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.771103 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wcf57\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.771242 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wcf57\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.771362 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvgvx\" (UniqueName: \"kubernetes.io/projected/bf19b11f-c43e-4f3b-a34c-53d3837705fb-kube-api-access-qvgvx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wcf57\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.776982 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wcf57\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.777028 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wcf57\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.784911 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wcf57\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.795788 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvgvx\" (UniqueName: \"kubernetes.io/projected/bf19b11f-c43e-4f3b-a34c-53d3837705fb-kube-api-access-qvgvx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wcf57\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:16 crc kubenswrapper[4744]: I1008 09:51:16.892526 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:51:17 crc kubenswrapper[4744]: I1008 09:51:17.392821 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57"] Oct 08 09:51:17 crc kubenswrapper[4744]: I1008 09:51:17.470957 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" event={"ID":"bf19b11f-c43e-4f3b-a34c-53d3837705fb","Type":"ContainerStarted","Data":"d3f22f23aaff38310c4f9b0a10f11a640ac4388aeb5aa738fceecfb4db7fba14"} Oct 08 09:51:18 crc kubenswrapper[4744]: I1008 09:51:18.480692 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" event={"ID":"bf19b11f-c43e-4f3b-a34c-53d3837705fb","Type":"ContainerStarted","Data":"621a5cf013f1104a4768ed7a1ac626b34afb8d08ae4f80abc1e585688bb64df0"} Oct 08 09:51:18 crc kubenswrapper[4744]: I1008 09:51:18.493872 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" podStartSLOduration=1.824728458 podStartE2EDuration="2.49385405s" podCreationTimestamp="2025-10-08 09:51:16 +0000 UTC" firstStartedPulling="2025-10-08 09:51:17.407738939 +0000 UTC m=+2372.655384198" lastFinishedPulling="2025-10-08 09:51:18.076864551 +0000 UTC m=+2373.324509790" observedRunningTime="2025-10-08 09:51:18.492839552 +0000 UTC m=+2373.740484801" watchObservedRunningTime="2025-10-08 09:51:18.49385405 +0000 UTC m=+2373.741499289" Oct 08 09:51:27 crc kubenswrapper[4744]: I1008 09:51:27.452706 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:51:27 crc kubenswrapper[4744]: E1008 09:51:27.453431 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:51:38 crc kubenswrapper[4744]: I1008 09:51:38.453641 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:51:38 crc kubenswrapper[4744]: E1008 09:51:38.454433 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:51:51 crc kubenswrapper[4744]: I1008 09:51:51.454361 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:51:51 crc kubenswrapper[4744]: E1008 09:51:51.456600 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:52:02 crc kubenswrapper[4744]: I1008 09:52:02.453488 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:52:02 crc kubenswrapper[4744]: E1008 09:52:02.454268 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:52:03 crc kubenswrapper[4744]: I1008 09:52:03.935864 4744 generic.go:334] "Generic (PLEG): container finished" podID="bf19b11f-c43e-4f3b-a34c-53d3837705fb" containerID="621a5cf013f1104a4768ed7a1ac626b34afb8d08ae4f80abc1e585688bb64df0" exitCode=0 Oct 08 09:52:03 crc kubenswrapper[4744]: I1008 09:52:03.935958 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" event={"ID":"bf19b11f-c43e-4f3b-a34c-53d3837705fb","Type":"ContainerDied","Data":"621a5cf013f1104a4768ed7a1ac626b34afb8d08ae4f80abc1e585688bb64df0"} Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.322122 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.507754 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-ssh-key\") pod \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.508209 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvgvx\" (UniqueName: \"kubernetes.io/projected/bf19b11f-c43e-4f3b-a34c-53d3837705fb-kube-api-access-qvgvx\") pod \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.508346 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-inventory\") pod \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.508510 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-ceph\") pod \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\" (UID: \"bf19b11f-c43e-4f3b-a34c-53d3837705fb\") " Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.513175 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-ceph" (OuterVolumeSpecName: "ceph") pod "bf19b11f-c43e-4f3b-a34c-53d3837705fb" (UID: "bf19b11f-c43e-4f3b-a34c-53d3837705fb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.517637 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf19b11f-c43e-4f3b-a34c-53d3837705fb-kube-api-access-qvgvx" (OuterVolumeSpecName: "kube-api-access-qvgvx") pod "bf19b11f-c43e-4f3b-a34c-53d3837705fb" (UID: "bf19b11f-c43e-4f3b-a34c-53d3837705fb"). InnerVolumeSpecName "kube-api-access-qvgvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.534145 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-inventory" (OuterVolumeSpecName: "inventory") pod "bf19b11f-c43e-4f3b-a34c-53d3837705fb" (UID: "bf19b11f-c43e-4f3b-a34c-53d3837705fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.538532 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf19b11f-c43e-4f3b-a34c-53d3837705fb" (UID: "bf19b11f-c43e-4f3b-a34c-53d3837705fb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.610753 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.610794 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvgvx\" (UniqueName: \"kubernetes.io/projected/bf19b11f-c43e-4f3b-a34c-53d3837705fb-kube-api-access-qvgvx\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.610806 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.610814 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf19b11f-c43e-4f3b-a34c-53d3837705fb-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.958068 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" event={"ID":"bf19b11f-c43e-4f3b-a34c-53d3837705fb","Type":"ContainerDied","Data":"d3f22f23aaff38310c4f9b0a10f11a640ac4388aeb5aa738fceecfb4db7fba14"} Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.958139 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3f22f23aaff38310c4f9b0a10f11a640ac4388aeb5aa738fceecfb4db7fba14" Oct 08 09:52:05 crc kubenswrapper[4744]: I1008 09:52:05.958638 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wcf57" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.422781 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-f6q5g"] Oct 08 09:52:06 crc kubenswrapper[4744]: E1008 09:52:06.423114 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf19b11f-c43e-4f3b-a34c-53d3837705fb" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.423126 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf19b11f-c43e-4f3b-a34c-53d3837705fb" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.423316 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf19b11f-c43e-4f3b-a34c-53d3837705fb" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.423917 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.426013 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.426363 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.426745 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.427196 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.435646 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.443473 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-f6q5g"] Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.526985 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-f6q5g\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.527048 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml98k\" (UniqueName: \"kubernetes.io/projected/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-kube-api-access-ml98k\") pod \"ssh-known-hosts-edpm-deployment-f6q5g\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.527090 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-ceph\") pod \"ssh-known-hosts-edpm-deployment-f6q5g\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.527124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-f6q5g\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.628881 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-f6q5g\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.628932 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml98k\" (UniqueName: \"kubernetes.io/projected/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-kube-api-access-ml98k\") pod \"ssh-known-hosts-edpm-deployment-f6q5g\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.628963 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-ceph\") pod \"ssh-known-hosts-edpm-deployment-f6q5g\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.628989 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-f6q5g\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.635119 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-ceph\") pod \"ssh-known-hosts-edpm-deployment-f6q5g\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.635132 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-f6q5g\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.636777 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-f6q5g\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.648282 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml98k\" (UniqueName: \"kubernetes.io/projected/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-kube-api-access-ml98k\") pod \"ssh-known-hosts-edpm-deployment-f6q5g\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:06 crc kubenswrapper[4744]: I1008 09:52:06.740794 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:07 crc kubenswrapper[4744]: I1008 09:52:07.262197 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-f6q5g"] Oct 08 09:52:07 crc kubenswrapper[4744]: I1008 09:52:07.974401 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" event={"ID":"ed43f66e-c9f8-4de0-9604-f5a1b993ed03","Type":"ContainerStarted","Data":"bca7db06635d6e85f92cde66620154a494124d6dcd703cb6c624e8c66ed9651f"} Oct 08 09:52:08 crc kubenswrapper[4744]: I1008 09:52:08.987639 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" event={"ID":"ed43f66e-c9f8-4de0-9604-f5a1b993ed03","Type":"ContainerStarted","Data":"2ece0a1fd5d58ed17d2dfcaadabc6d65b711a3e9cc950f43d4a2d9b9431d562c"} Oct 08 09:52:09 crc kubenswrapper[4744]: I1008 09:52:09.012212 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" podStartSLOduration=2.190806071 podStartE2EDuration="3.012193747s" podCreationTimestamp="2025-10-08 09:52:06 +0000 UTC" firstStartedPulling="2025-10-08 09:52:07.273785113 +0000 UTC m=+2422.521430352" lastFinishedPulling="2025-10-08 09:52:08.095172789 +0000 UTC m=+2423.342818028" observedRunningTime="2025-10-08 09:52:09.007739143 +0000 UTC m=+2424.255384392" watchObservedRunningTime="2025-10-08 09:52:09.012193747 +0000 UTC m=+2424.259838986" Oct 08 09:52:16 crc kubenswrapper[4744]: I1008 09:52:16.455676 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:52:16 crc kubenswrapper[4744]: E1008 09:52:16.456485 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:52:18 crc kubenswrapper[4744]: I1008 09:52:18.070316 4744 generic.go:334] "Generic (PLEG): container finished" podID="ed43f66e-c9f8-4de0-9604-f5a1b993ed03" containerID="2ece0a1fd5d58ed17d2dfcaadabc6d65b711a3e9cc950f43d4a2d9b9431d562c" exitCode=0 Oct 08 09:52:18 crc kubenswrapper[4744]: I1008 09:52:18.070392 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" event={"ID":"ed43f66e-c9f8-4de0-9604-f5a1b993ed03","Type":"ContainerDied","Data":"2ece0a1fd5d58ed17d2dfcaadabc6d65b711a3e9cc950f43d4a2d9b9431d562c"} Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.471634 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.653195 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ml98k\" (UniqueName: \"kubernetes.io/projected/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-kube-api-access-ml98k\") pod \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.653834 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-ceph\") pod \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.654058 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-inventory-0\") pod \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.654179 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-ssh-key-openstack-edpm-ipam\") pod \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\" (UID: \"ed43f66e-c9f8-4de0-9604-f5a1b993ed03\") " Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.660652 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-kube-api-access-ml98k" (OuterVolumeSpecName: "kube-api-access-ml98k") pod "ed43f66e-c9f8-4de0-9604-f5a1b993ed03" (UID: "ed43f66e-c9f8-4de0-9604-f5a1b993ed03"). InnerVolumeSpecName "kube-api-access-ml98k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.662228 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-ceph" (OuterVolumeSpecName: "ceph") pod "ed43f66e-c9f8-4de0-9604-f5a1b993ed03" (UID: "ed43f66e-c9f8-4de0-9604-f5a1b993ed03"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.683146 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "ed43f66e-c9f8-4de0-9604-f5a1b993ed03" (UID: "ed43f66e-c9f8-4de0-9604-f5a1b993ed03"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.683763 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ed43f66e-c9f8-4de0-9604-f5a1b993ed03" (UID: "ed43f66e-c9f8-4de0-9604-f5a1b993ed03"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.756869 4744 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.756904 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.756920 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ml98k\" (UniqueName: \"kubernetes.io/projected/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-kube-api-access-ml98k\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:19 crc kubenswrapper[4744]: I1008 09:52:19.756932 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ed43f66e-c9f8-4de0-9604-f5a1b993ed03-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.088401 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" event={"ID":"ed43f66e-c9f8-4de0-9604-f5a1b993ed03","Type":"ContainerDied","Data":"bca7db06635d6e85f92cde66620154a494124d6dcd703cb6c624e8c66ed9651f"} Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.088450 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bca7db06635d6e85f92cde66620154a494124d6dcd703cb6c624e8c66ed9651f" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.088516 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-f6q5g" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.180094 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg"] Oct 08 09:52:20 crc kubenswrapper[4744]: E1008 09:52:20.180859 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed43f66e-c9f8-4de0-9604-f5a1b993ed03" containerName="ssh-known-hosts-edpm-deployment" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.180883 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed43f66e-c9f8-4de0-9604-f5a1b993ed03" containerName="ssh-known-hosts-edpm-deployment" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.181070 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed43f66e-c9f8-4de0-9604-f5a1b993ed03" containerName="ssh-known-hosts-edpm-deployment" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.181629 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.183810 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.183919 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.184218 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.184245 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.190072 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.192256 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg"] Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.367355 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmbgg\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.367449 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmbgg\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.367512 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmbgg\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.367912 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66gnn\" (UniqueName: \"kubernetes.io/projected/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-kube-api-access-66gnn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmbgg\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.469818 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmbgg\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.470316 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmbgg\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.471327 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66gnn\" (UniqueName: \"kubernetes.io/projected/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-kube-api-access-66gnn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmbgg\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.471746 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmbgg\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.477331 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmbgg\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.479060 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmbgg\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.480360 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmbgg\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.490923 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66gnn\" (UniqueName: \"kubernetes.io/projected/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-kube-api-access-66gnn\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmbgg\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.499813 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:20 crc kubenswrapper[4744]: I1008 09:52:20.822575 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg"] Oct 08 09:52:21 crc kubenswrapper[4744]: I1008 09:52:21.098040 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" event={"ID":"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0","Type":"ContainerStarted","Data":"8e278ece8a9d552dd9262d461b62725cf0c5b6ecd21461ccfbed7a66d6ab66b8"} Oct 08 09:52:22 crc kubenswrapper[4744]: I1008 09:52:22.113485 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" event={"ID":"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0","Type":"ContainerStarted","Data":"61cfdb4d863efa8e3ceaceffec5248c304f97a259cf5f0024567a9b3be204b10"} Oct 08 09:52:22 crc kubenswrapper[4744]: I1008 09:52:22.137087 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" podStartSLOduration=1.55907833 podStartE2EDuration="2.137064806s" podCreationTimestamp="2025-10-08 09:52:20 +0000 UTC" firstStartedPulling="2025-10-08 09:52:20.834721932 +0000 UTC m=+2436.082367171" lastFinishedPulling="2025-10-08 09:52:21.412708408 +0000 UTC m=+2436.660353647" observedRunningTime="2025-10-08 09:52:22.132650063 +0000 UTC m=+2437.380295302" watchObservedRunningTime="2025-10-08 09:52:22.137064806 +0000 UTC m=+2437.384710045" Oct 08 09:52:28 crc kubenswrapper[4744]: I1008 09:52:28.453766 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:52:28 crc kubenswrapper[4744]: E1008 09:52:28.454974 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:52:30 crc kubenswrapper[4744]: I1008 09:52:30.190058 4744 generic.go:334] "Generic (PLEG): container finished" podID="77eb2e7f-5360-4589-bb9f-56d6a28fc3b0" containerID="61cfdb4d863efa8e3ceaceffec5248c304f97a259cf5f0024567a9b3be204b10" exitCode=0 Oct 08 09:52:30 crc kubenswrapper[4744]: I1008 09:52:30.190173 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" event={"ID":"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0","Type":"ContainerDied","Data":"61cfdb4d863efa8e3ceaceffec5248c304f97a259cf5f0024567a9b3be204b10"} Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.653532 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.839828 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66gnn\" (UniqueName: \"kubernetes.io/projected/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-kube-api-access-66gnn\") pod \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.839900 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-ceph\") pod \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.839932 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-ssh-key\") pod \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.840127 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-inventory\") pod \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\" (UID: \"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0\") " Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.845228 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-ceph" (OuterVolumeSpecName: "ceph") pod "77eb2e7f-5360-4589-bb9f-56d6a28fc3b0" (UID: "77eb2e7f-5360-4589-bb9f-56d6a28fc3b0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.845714 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-kube-api-access-66gnn" (OuterVolumeSpecName: "kube-api-access-66gnn") pod "77eb2e7f-5360-4589-bb9f-56d6a28fc3b0" (UID: "77eb2e7f-5360-4589-bb9f-56d6a28fc3b0"). InnerVolumeSpecName "kube-api-access-66gnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.873332 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-inventory" (OuterVolumeSpecName: "inventory") pod "77eb2e7f-5360-4589-bb9f-56d6a28fc3b0" (UID: "77eb2e7f-5360-4589-bb9f-56d6a28fc3b0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.873758 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "77eb2e7f-5360-4589-bb9f-56d6a28fc3b0" (UID: "77eb2e7f-5360-4589-bb9f-56d6a28fc3b0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.941529 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.941557 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.941568 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:31 crc kubenswrapper[4744]: I1008 09:52:31.941579 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66gnn\" (UniqueName: \"kubernetes.io/projected/77eb2e7f-5360-4589-bb9f-56d6a28fc3b0-kube-api-access-66gnn\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.218076 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" event={"ID":"77eb2e7f-5360-4589-bb9f-56d6a28fc3b0","Type":"ContainerDied","Data":"8e278ece8a9d552dd9262d461b62725cf0c5b6ecd21461ccfbed7a66d6ab66b8"} Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.218115 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e278ece8a9d552dd9262d461b62725cf0c5b6ecd21461ccfbed7a66d6ab66b8" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.218171 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmbgg" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.297785 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2"] Oct 08 09:52:32 crc kubenswrapper[4744]: E1008 09:52:32.298223 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77eb2e7f-5360-4589-bb9f-56d6a28fc3b0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.298246 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="77eb2e7f-5360-4589-bb9f-56d6a28fc3b0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.299327 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="77eb2e7f-5360-4589-bb9f-56d6a28fc3b0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.300422 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.306901 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.306947 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.307141 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.307308 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.307465 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.316402 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2"] Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.451201 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.451262 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.451286 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z5bg\" (UniqueName: \"kubernetes.io/projected/446dc944-bd86-4c7a-a9da-bc1193d34d33-kube-api-access-4z5bg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.451435 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.552501 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.552740 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.552799 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.552836 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z5bg\" (UniqueName: \"kubernetes.io/projected/446dc944-bd86-4c7a-a9da-bc1193d34d33-kube-api-access-4z5bg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.557084 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.557513 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.559957 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.570631 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z5bg\" (UniqueName: \"kubernetes.io/projected/446dc944-bd86-4c7a-a9da-bc1193d34d33-kube-api-access-4z5bg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:32 crc kubenswrapper[4744]: I1008 09:52:32.622846 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:33 crc kubenswrapper[4744]: I1008 09:52:33.146914 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2"] Oct 08 09:52:33 crc kubenswrapper[4744]: I1008 09:52:33.227756 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" event={"ID":"446dc944-bd86-4c7a-a9da-bc1193d34d33","Type":"ContainerStarted","Data":"864ee4792f296ab4879a12bfd7c04577d7410c47260bcbf93a7da34e0c4ea083"} Oct 08 09:52:34 crc kubenswrapper[4744]: I1008 09:52:34.237576 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" event={"ID":"446dc944-bd86-4c7a-a9da-bc1193d34d33","Type":"ContainerStarted","Data":"339e61c7e6999c4f5b290d0536859806e78fcd0f82b91b47047bee1cdb7efd66"} Oct 08 09:52:43 crc kubenswrapper[4744]: I1008 09:52:43.453139 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:52:43 crc kubenswrapper[4744]: E1008 09:52:43.454016 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:52:44 crc kubenswrapper[4744]: I1008 09:52:44.327947 4744 generic.go:334] "Generic (PLEG): container finished" podID="446dc944-bd86-4c7a-a9da-bc1193d34d33" containerID="339e61c7e6999c4f5b290d0536859806e78fcd0f82b91b47047bee1cdb7efd66" exitCode=0 Oct 08 09:52:44 crc kubenswrapper[4744]: I1008 09:52:44.328155 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" event={"ID":"446dc944-bd86-4c7a-a9da-bc1193d34d33","Type":"ContainerDied","Data":"339e61c7e6999c4f5b290d0536859806e78fcd0f82b91b47047bee1cdb7efd66"} Oct 08 09:52:45 crc kubenswrapper[4744]: I1008 09:52:45.742943 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:45 crc kubenswrapper[4744]: I1008 09:52:45.901659 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-ceph\") pod \"446dc944-bd86-4c7a-a9da-bc1193d34d33\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " Oct 08 09:52:45 crc kubenswrapper[4744]: I1008 09:52:45.901759 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-inventory\") pod \"446dc944-bd86-4c7a-a9da-bc1193d34d33\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " Oct 08 09:52:45 crc kubenswrapper[4744]: I1008 09:52:45.901828 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z5bg\" (UniqueName: \"kubernetes.io/projected/446dc944-bd86-4c7a-a9da-bc1193d34d33-kube-api-access-4z5bg\") pod \"446dc944-bd86-4c7a-a9da-bc1193d34d33\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " Oct 08 09:52:45 crc kubenswrapper[4744]: I1008 09:52:45.901882 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-ssh-key\") pod \"446dc944-bd86-4c7a-a9da-bc1193d34d33\" (UID: \"446dc944-bd86-4c7a-a9da-bc1193d34d33\") " Oct 08 09:52:45 crc kubenswrapper[4744]: I1008 09:52:45.911794 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-ceph" (OuterVolumeSpecName: "ceph") pod "446dc944-bd86-4c7a-a9da-bc1193d34d33" (UID: "446dc944-bd86-4c7a-a9da-bc1193d34d33"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:52:45 crc kubenswrapper[4744]: I1008 09:52:45.911860 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/446dc944-bd86-4c7a-a9da-bc1193d34d33-kube-api-access-4z5bg" (OuterVolumeSpecName: "kube-api-access-4z5bg") pod "446dc944-bd86-4c7a-a9da-bc1193d34d33" (UID: "446dc944-bd86-4c7a-a9da-bc1193d34d33"). InnerVolumeSpecName "kube-api-access-4z5bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:52:45 crc kubenswrapper[4744]: I1008 09:52:45.927738 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "446dc944-bd86-4c7a-a9da-bc1193d34d33" (UID: "446dc944-bd86-4c7a-a9da-bc1193d34d33"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:52:45 crc kubenswrapper[4744]: I1008 09:52:45.930340 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-inventory" (OuterVolumeSpecName: "inventory") pod "446dc944-bd86-4c7a-a9da-bc1193d34d33" (UID: "446dc944-bd86-4c7a-a9da-bc1193d34d33"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.004518 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.004552 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z5bg\" (UniqueName: \"kubernetes.io/projected/446dc944-bd86-4c7a-a9da-bc1193d34d33-kube-api-access-4z5bg\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.004563 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.004571 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/446dc944-bd86-4c7a-a9da-bc1193d34d33-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.356036 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" event={"ID":"446dc944-bd86-4c7a-a9da-bc1193d34d33","Type":"ContainerDied","Data":"864ee4792f296ab4879a12bfd7c04577d7410c47260bcbf93a7da34e0c4ea083"} Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.356356 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="864ee4792f296ab4879a12bfd7c04577d7410c47260bcbf93a7da34e0c4ea083" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.356088 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.463917 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd"] Oct 08 09:52:46 crc kubenswrapper[4744]: E1008 09:52:46.464592 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446dc944-bd86-4c7a-a9da-bc1193d34d33" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.464688 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="446dc944-bd86-4c7a-a9da-bc1193d34d33" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.465166 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="446dc944-bd86-4c7a-a9da-bc1193d34d33" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.466073 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.468303 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.468545 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.468723 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.470640 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.470772 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.470915 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.471059 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.473310 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.506008 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd"] Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.512957 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.513022 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.513054 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.513126 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.513165 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.513197 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.513248 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.513281 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb47x\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-kube-api-access-hb47x\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.513442 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.513513 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.513561 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.513698 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.513773 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.615567 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.615638 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.615694 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.616827 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.616924 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb47x\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-kube-api-access-hb47x\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.616975 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.617006 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.617040 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.617118 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.617179 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.617212 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.617246 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.617278 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.621663 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.622004 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.624968 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.627173 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.627792 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.628304 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.629242 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.630324 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.630606 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.631436 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.633958 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.634702 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb47x\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-kube-api-access-hb47x\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.637654 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bctfd\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:46 crc kubenswrapper[4744]: I1008 09:52:46.788677 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:52:47 crc kubenswrapper[4744]: I1008 09:52:47.118892 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd"] Oct 08 09:52:47 crc kubenswrapper[4744]: I1008 09:52:47.367797 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" event={"ID":"3876b90a-308b-4320-a8ab-f81a48295c1f","Type":"ContainerStarted","Data":"15e7a149887e2753335638be3c3920104f3f7e721b1e160ed22271bb3dba2e75"} Oct 08 09:52:48 crc kubenswrapper[4744]: I1008 09:52:48.378251 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" event={"ID":"3876b90a-308b-4320-a8ab-f81a48295c1f","Type":"ContainerStarted","Data":"cd7c56c2118e3c5ecef3ce07ce4e2290114f5b7c221fbc4f53cd69d985a139be"} Oct 08 09:52:48 crc kubenswrapper[4744]: I1008 09:52:48.402252 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" podStartSLOduration=1.898128239 podStartE2EDuration="2.402236496s" podCreationTimestamp="2025-10-08 09:52:46 +0000 UTC" firstStartedPulling="2025-10-08 09:52:47.123806934 +0000 UTC m=+2462.371452173" lastFinishedPulling="2025-10-08 09:52:47.627915191 +0000 UTC m=+2462.875560430" observedRunningTime="2025-10-08 09:52:48.396772164 +0000 UTC m=+2463.644417403" watchObservedRunningTime="2025-10-08 09:52:48.402236496 +0000 UTC m=+2463.649881735" Oct 08 09:52:54 crc kubenswrapper[4744]: I1008 09:52:54.453194 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:52:54 crc kubenswrapper[4744]: E1008 09:52:54.453847 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:53:08 crc kubenswrapper[4744]: I1008 09:53:08.453160 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:53:08 crc kubenswrapper[4744]: E1008 09:53:08.454034 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:53:18 crc kubenswrapper[4744]: I1008 09:53:18.623066 4744 generic.go:334] "Generic (PLEG): container finished" podID="3876b90a-308b-4320-a8ab-f81a48295c1f" containerID="cd7c56c2118e3c5ecef3ce07ce4e2290114f5b7c221fbc4f53cd69d985a139be" exitCode=0 Oct 08 09:53:18 crc kubenswrapper[4744]: I1008 09:53:18.623120 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" event={"ID":"3876b90a-308b-4320-a8ab-f81a48295c1f","Type":"ContainerDied","Data":"cd7c56c2118e3c5ecef3ce07ce4e2290114f5b7c221fbc4f53cd69d985a139be"} Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.040876 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.112863 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-inventory\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.112957 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.113026 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.113057 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-neutron-metadata-combined-ca-bundle\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.113979 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-libvirt-combined-ca-bundle\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.114002 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ceph\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.114024 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-bootstrap-combined-ca-bundle\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.114052 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ovn-combined-ca-bundle\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.114075 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ssh-key\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.114108 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-repo-setup-combined-ca-bundle\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.114141 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hb47x\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-kube-api-access-hb47x\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.114168 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.114219 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-nova-combined-ca-bundle\") pod \"3876b90a-308b-4320-a8ab-f81a48295c1f\" (UID: \"3876b90a-308b-4320-a8ab-f81a48295c1f\") " Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.118304 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.118635 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ceph" (OuterVolumeSpecName: "ceph") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.119132 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.119271 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.121775 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.122128 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.122538 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.123118 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-kube-api-access-hb47x" (OuterVolumeSpecName: "kube-api-access-hb47x") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "kube-api-access-hb47x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.123536 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.125831 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.131545 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.146503 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.146898 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-inventory" (OuterVolumeSpecName: "inventory") pod "3876b90a-308b-4320-a8ab-f81a48295c1f" (UID: "3876b90a-308b-4320-a8ab-f81a48295c1f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216400 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216434 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216446 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216458 4744 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216470 4744 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216478 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216486 4744 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216495 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216503 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216511 4744 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216520 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hb47x\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-kube-api-access-hb47x\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216528 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3876b90a-308b-4320-a8ab-f81a48295c1f-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.216536 4744 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3876b90a-308b-4320-a8ab-f81a48295c1f-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.453538 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:53:20 crc kubenswrapper[4744]: E1008 09:53:20.453901 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.640915 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" event={"ID":"3876b90a-308b-4320-a8ab-f81a48295c1f","Type":"ContainerDied","Data":"15e7a149887e2753335638be3c3920104f3f7e721b1e160ed22271bb3dba2e75"} Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.640952 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15e7a149887e2753335638be3c3920104f3f7e721b1e160ed22271bb3dba2e75" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.641119 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bctfd" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.759883 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k"] Oct 08 09:53:20 crc kubenswrapper[4744]: E1008 09:53:20.760220 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3876b90a-308b-4320-a8ab-f81a48295c1f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.760237 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3876b90a-308b-4320-a8ab-f81a48295c1f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.760439 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3876b90a-308b-4320-a8ab-f81a48295c1f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.761033 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.763526 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.763567 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.763780 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.763940 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.764093 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.775928 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k"] Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.930017 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.930066 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.930094 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:20 crc kubenswrapper[4744]: I1008 09:53:20.930144 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krgl4\" (UniqueName: \"kubernetes.io/projected/664b4c04-3a69-444f-aa77-621229ee5b96-kube-api-access-krgl4\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:21 crc kubenswrapper[4744]: I1008 09:53:21.031465 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:21 crc kubenswrapper[4744]: I1008 09:53:21.031809 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:21 crc kubenswrapper[4744]: I1008 09:53:21.031862 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:21 crc kubenswrapper[4744]: I1008 09:53:21.031893 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krgl4\" (UniqueName: \"kubernetes.io/projected/664b4c04-3a69-444f-aa77-621229ee5b96-kube-api-access-krgl4\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:21 crc kubenswrapper[4744]: I1008 09:53:21.034744 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:21 crc kubenswrapper[4744]: I1008 09:53:21.036877 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:21 crc kubenswrapper[4744]: I1008 09:53:21.044442 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:21 crc kubenswrapper[4744]: I1008 09:53:21.052628 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krgl4\" (UniqueName: \"kubernetes.io/projected/664b4c04-3a69-444f-aa77-621229ee5b96-kube-api-access-krgl4\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:21 crc kubenswrapper[4744]: I1008 09:53:21.077624 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:21 crc kubenswrapper[4744]: I1008 09:53:21.569039 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k"] Oct 08 09:53:21 crc kubenswrapper[4744]: I1008 09:53:21.648524 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" event={"ID":"664b4c04-3a69-444f-aa77-621229ee5b96","Type":"ContainerStarted","Data":"4cb2b7aaf67c48370f320e14505679494fa9f17f6341fe25224231a7046f14f9"} Oct 08 09:53:22 crc kubenswrapper[4744]: I1008 09:53:22.659399 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" event={"ID":"664b4c04-3a69-444f-aa77-621229ee5b96","Type":"ContainerStarted","Data":"2118ff2124e742aac64f9a40b163ddfddf214a8f9587029e51fbee544fa79ed4"} Oct 08 09:53:22 crc kubenswrapper[4744]: I1008 09:53:22.696558 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" podStartSLOduration=2.134632904 podStartE2EDuration="2.696357164s" podCreationTimestamp="2025-10-08 09:53:20 +0000 UTC" firstStartedPulling="2025-10-08 09:53:21.571248558 +0000 UTC m=+2496.818893797" lastFinishedPulling="2025-10-08 09:53:22.132972808 +0000 UTC m=+2497.380618057" observedRunningTime="2025-10-08 09:53:22.678484597 +0000 UTC m=+2497.926129846" watchObservedRunningTime="2025-10-08 09:53:22.696357164 +0000 UTC m=+2497.944002403" Oct 08 09:53:27 crc kubenswrapper[4744]: I1008 09:53:27.709878 4744 generic.go:334] "Generic (PLEG): container finished" podID="664b4c04-3a69-444f-aa77-621229ee5b96" containerID="2118ff2124e742aac64f9a40b163ddfddf214a8f9587029e51fbee544fa79ed4" exitCode=0 Oct 08 09:53:27 crc kubenswrapper[4744]: I1008 09:53:27.709945 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" event={"ID":"664b4c04-3a69-444f-aa77-621229ee5b96","Type":"ContainerDied","Data":"2118ff2124e742aac64f9a40b163ddfddf214a8f9587029e51fbee544fa79ed4"} Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.177692 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.289519 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-ssh-key\") pod \"664b4c04-3a69-444f-aa77-621229ee5b96\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.289692 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krgl4\" (UniqueName: \"kubernetes.io/projected/664b4c04-3a69-444f-aa77-621229ee5b96-kube-api-access-krgl4\") pod \"664b4c04-3a69-444f-aa77-621229ee5b96\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.289785 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-inventory\") pod \"664b4c04-3a69-444f-aa77-621229ee5b96\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.289853 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-ceph\") pod \"664b4c04-3a69-444f-aa77-621229ee5b96\" (UID: \"664b4c04-3a69-444f-aa77-621229ee5b96\") " Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.295352 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-ceph" (OuterVolumeSpecName: "ceph") pod "664b4c04-3a69-444f-aa77-621229ee5b96" (UID: "664b4c04-3a69-444f-aa77-621229ee5b96"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.299592 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/664b4c04-3a69-444f-aa77-621229ee5b96-kube-api-access-krgl4" (OuterVolumeSpecName: "kube-api-access-krgl4") pod "664b4c04-3a69-444f-aa77-621229ee5b96" (UID: "664b4c04-3a69-444f-aa77-621229ee5b96"). InnerVolumeSpecName "kube-api-access-krgl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.317769 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "664b4c04-3a69-444f-aa77-621229ee5b96" (UID: "664b4c04-3a69-444f-aa77-621229ee5b96"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.322624 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-inventory" (OuterVolumeSpecName: "inventory") pod "664b4c04-3a69-444f-aa77-621229ee5b96" (UID: "664b4c04-3a69-444f-aa77-621229ee5b96"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.391968 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.392000 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.392009 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/664b4c04-3a69-444f-aa77-621229ee5b96-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.392019 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krgl4\" (UniqueName: \"kubernetes.io/projected/664b4c04-3a69-444f-aa77-621229ee5b96-kube-api-access-krgl4\") on node \"crc\" DevicePath \"\"" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.762030 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" event={"ID":"664b4c04-3a69-444f-aa77-621229ee5b96","Type":"ContainerDied","Data":"4cb2b7aaf67c48370f320e14505679494fa9f17f6341fe25224231a7046f14f9"} Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.762104 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4cb2b7aaf67c48370f320e14505679494fa9f17f6341fe25224231a7046f14f9" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.762214 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.840640 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4"] Oct 08 09:53:29 crc kubenswrapper[4744]: E1008 09:53:29.840964 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="664b4c04-3a69-444f-aa77-621229ee5b96" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.840978 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="664b4c04-3a69-444f-aa77-621229ee5b96" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.841150 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="664b4c04-3a69-444f-aa77-621229ee5b96" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.841726 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.844113 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.844823 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.844863 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.844920 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.845518 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.846523 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.867716 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4"] Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.904392 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.904483 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.904554 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.904841 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwrc7\" (UniqueName: \"kubernetes.io/projected/fa1452ca-c8a1-4185-b5ee-9119851d2e20-kube-api-access-dwrc7\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.904887 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:29 crc kubenswrapper[4744]: I1008 09:53:29.904968 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.006389 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwrc7\" (UniqueName: \"kubernetes.io/projected/fa1452ca-c8a1-4185-b5ee-9119851d2e20-kube-api-access-dwrc7\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.006427 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.006479 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.006520 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.006555 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.006611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.009720 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.011288 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.013968 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.022884 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.024885 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.028670 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwrc7\" (UniqueName: \"kubernetes.io/projected/fa1452ca-c8a1-4185-b5ee-9119851d2e20-kube-api-access-dwrc7\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-qslf4\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.160623 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.740387 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4"] Oct 08 09:53:30 crc kubenswrapper[4744]: I1008 09:53:30.782433 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" event={"ID":"fa1452ca-c8a1-4185-b5ee-9119851d2e20","Type":"ContainerStarted","Data":"aced93b815045c52a1a7f46a6c9d60a777e047818fa135f907e3af3702ca3269"} Oct 08 09:53:31 crc kubenswrapper[4744]: I1008 09:53:31.792893 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" event={"ID":"fa1452ca-c8a1-4185-b5ee-9119851d2e20","Type":"ContainerStarted","Data":"88dcfdedae8925a4781e24ebb1e27ae1107268aa694997d70779aaf84b4a070e"} Oct 08 09:53:31 crc kubenswrapper[4744]: I1008 09:53:31.815596 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" podStartSLOduration=2.181034388 podStartE2EDuration="2.815578875s" podCreationTimestamp="2025-10-08 09:53:29 +0000 UTC" firstStartedPulling="2025-10-08 09:53:30.751010683 +0000 UTC m=+2505.998655932" lastFinishedPulling="2025-10-08 09:53:31.38555517 +0000 UTC m=+2506.633200419" observedRunningTime="2025-10-08 09:53:31.807750877 +0000 UTC m=+2507.055396116" watchObservedRunningTime="2025-10-08 09:53:31.815578875 +0000 UTC m=+2507.063224114" Oct 08 09:53:32 crc kubenswrapper[4744]: I1008 09:53:32.454442 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:53:32 crc kubenswrapper[4744]: E1008 09:53:32.454673 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:53:45 crc kubenswrapper[4744]: I1008 09:53:45.460426 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:53:45 crc kubenswrapper[4744]: E1008 09:53:45.461584 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:53:56 crc kubenswrapper[4744]: I1008 09:53:56.468790 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:53:56 crc kubenswrapper[4744]: E1008 09:53:56.469554 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:54:11 crc kubenswrapper[4744]: I1008 09:54:11.454748 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:54:11 crc kubenswrapper[4744]: E1008 09:54:11.455743 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:54:22 crc kubenswrapper[4744]: I1008 09:54:22.452672 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:54:22 crc kubenswrapper[4744]: E1008 09:54:22.453341 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:54:33 crc kubenswrapper[4744]: I1008 09:54:33.453629 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:54:33 crc kubenswrapper[4744]: E1008 09:54:33.454332 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:54:46 crc kubenswrapper[4744]: I1008 09:54:46.414650 4744 generic.go:334] "Generic (PLEG): container finished" podID="fa1452ca-c8a1-4185-b5ee-9119851d2e20" containerID="88dcfdedae8925a4781e24ebb1e27ae1107268aa694997d70779aaf84b4a070e" exitCode=0 Oct 08 09:54:46 crc kubenswrapper[4744]: I1008 09:54:46.414908 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" event={"ID":"fa1452ca-c8a1-4185-b5ee-9119851d2e20","Type":"ContainerDied","Data":"88dcfdedae8925a4781e24ebb1e27ae1107268aa694997d70779aaf84b4a070e"} Oct 08 09:54:47 crc kubenswrapper[4744]: I1008 09:54:47.844977 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.016685 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ovncontroller-config-0\") pod \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.016755 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ovn-combined-ca-bundle\") pod \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.016823 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-inventory\") pod \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.016870 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwrc7\" (UniqueName: \"kubernetes.io/projected/fa1452ca-c8a1-4185-b5ee-9119851d2e20-kube-api-access-dwrc7\") pod \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.016888 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ssh-key\") pod \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.016929 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ceph\") pod \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\" (UID: \"fa1452ca-c8a1-4185-b5ee-9119851d2e20\") " Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.022450 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ceph" (OuterVolumeSpecName: "ceph") pod "fa1452ca-c8a1-4185-b5ee-9119851d2e20" (UID: "fa1452ca-c8a1-4185-b5ee-9119851d2e20"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.024539 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "fa1452ca-c8a1-4185-b5ee-9119851d2e20" (UID: "fa1452ca-c8a1-4185-b5ee-9119851d2e20"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.032723 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa1452ca-c8a1-4185-b5ee-9119851d2e20-kube-api-access-dwrc7" (OuterVolumeSpecName: "kube-api-access-dwrc7") pod "fa1452ca-c8a1-4185-b5ee-9119851d2e20" (UID: "fa1452ca-c8a1-4185-b5ee-9119851d2e20"). InnerVolumeSpecName "kube-api-access-dwrc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.055553 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-inventory" (OuterVolumeSpecName: "inventory") pod "fa1452ca-c8a1-4185-b5ee-9119851d2e20" (UID: "fa1452ca-c8a1-4185-b5ee-9119851d2e20"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.060256 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "fa1452ca-c8a1-4185-b5ee-9119851d2e20" (UID: "fa1452ca-c8a1-4185-b5ee-9119851d2e20"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.077170 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fa1452ca-c8a1-4185-b5ee-9119851d2e20" (UID: "fa1452ca-c8a1-4185-b5ee-9119851d2e20"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.118940 4744 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.118973 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.118983 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.118991 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwrc7\" (UniqueName: \"kubernetes.io/projected/fa1452ca-c8a1-4185-b5ee-9119851d2e20-kube-api-access-dwrc7\") on node \"crc\" DevicePath \"\"" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.118999 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.119007 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fa1452ca-c8a1-4185-b5ee-9119851d2e20-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.435196 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" event={"ID":"fa1452ca-c8a1-4185-b5ee-9119851d2e20","Type":"ContainerDied","Data":"aced93b815045c52a1a7f46a6c9d60a777e047818fa135f907e3af3702ca3269"} Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.435236 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aced93b815045c52a1a7f46a6c9d60a777e047818fa135f907e3af3702ca3269" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.435298 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-qslf4" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.453927 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:54:48 crc kubenswrapper[4744]: E1008 09:54:48.454134 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.511692 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx"] Oct 08 09:54:48 crc kubenswrapper[4744]: E1008 09:54:48.512067 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa1452ca-c8a1-4185-b5ee-9119851d2e20" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.512082 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa1452ca-c8a1-4185-b5ee-9119851d2e20" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.512256 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa1452ca-c8a1-4185-b5ee-9119851d2e20" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.512868 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.515627 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.515767 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.517571 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.517708 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.517752 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.517877 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.519028 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.526627 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx"] Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.630056 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.630193 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.630233 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.630467 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.630517 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxwjl\" (UniqueName: \"kubernetes.io/projected/3f2cc21c-f2e1-4775-8563-c794073928f4-kube-api-access-zxwjl\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.630578 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.630623 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.732024 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.732093 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.732119 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.732182 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.732207 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxwjl\" (UniqueName: \"kubernetes.io/projected/3f2cc21c-f2e1-4775-8563-c794073928f4-kube-api-access-zxwjl\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.732243 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.732261 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.735885 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.736897 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.736948 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.741079 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.752077 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.765000 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.775128 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxwjl\" (UniqueName: \"kubernetes.io/projected/3f2cc21c-f2e1-4775-8563-c794073928f4-kube-api-access-zxwjl\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:48 crc kubenswrapper[4744]: I1008 09:54:48.830959 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:54:49 crc kubenswrapper[4744]: I1008 09:54:49.494610 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 09:54:49 crc kubenswrapper[4744]: I1008 09:54:49.502299 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx"] Oct 08 09:54:50 crc kubenswrapper[4744]: I1008 09:54:50.451247 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" event={"ID":"3f2cc21c-f2e1-4775-8563-c794073928f4","Type":"ContainerStarted","Data":"a9a22fd10bf3b4dd058539065baae01a8000019115375d154c502b00ba325e17"} Oct 08 09:54:50 crc kubenswrapper[4744]: I1008 09:54:50.451576 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" event={"ID":"3f2cc21c-f2e1-4775-8563-c794073928f4","Type":"ContainerStarted","Data":"d77535d18c078710fc418470480735261bc35c1a27db1ce890ddc05a2cabf223"} Oct 08 09:54:50 crc kubenswrapper[4744]: I1008 09:54:50.478599 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" podStartSLOduration=1.91077645 podStartE2EDuration="2.478583289s" podCreationTimestamp="2025-10-08 09:54:48 +0000 UTC" firstStartedPulling="2025-10-08 09:54:49.494333092 +0000 UTC m=+2584.741978331" lastFinishedPulling="2025-10-08 09:54:50.062139941 +0000 UTC m=+2585.309785170" observedRunningTime="2025-10-08 09:54:50.473347733 +0000 UTC m=+2585.720992972" watchObservedRunningTime="2025-10-08 09:54:50.478583289 +0000 UTC m=+2585.726228528" Oct 08 09:55:02 crc kubenswrapper[4744]: I1008 09:55:02.453069 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 09:55:03 crc kubenswrapper[4744]: I1008 09:55:03.552560 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"aa495b33f33b23e8dcca10511302304051a76ce3efd1ef1a213237865410ba57"} Oct 08 09:55:38 crc kubenswrapper[4744]: I1008 09:55:38.766942 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4wsmr"] Oct 08 09:55:38 crc kubenswrapper[4744]: I1008 09:55:38.769929 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:38 crc kubenswrapper[4744]: I1008 09:55:38.784159 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4wsmr"] Oct 08 09:55:38 crc kubenswrapper[4744]: I1008 09:55:38.915629 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-utilities\") pod \"redhat-operators-4wsmr\" (UID: \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\") " pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:38 crc kubenswrapper[4744]: I1008 09:55:38.916476 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdchb\" (UniqueName: \"kubernetes.io/projected/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-kube-api-access-zdchb\") pod \"redhat-operators-4wsmr\" (UID: \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\") " pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:38 crc kubenswrapper[4744]: I1008 09:55:38.916587 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-catalog-content\") pod \"redhat-operators-4wsmr\" (UID: \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\") " pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:39 crc kubenswrapper[4744]: I1008 09:55:39.018931 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdchb\" (UniqueName: \"kubernetes.io/projected/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-kube-api-access-zdchb\") pod \"redhat-operators-4wsmr\" (UID: \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\") " pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:39 crc kubenswrapper[4744]: I1008 09:55:39.019285 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-catalog-content\") pod \"redhat-operators-4wsmr\" (UID: \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\") " pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:39 crc kubenswrapper[4744]: I1008 09:55:39.019718 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-catalog-content\") pod \"redhat-operators-4wsmr\" (UID: \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\") " pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:39 crc kubenswrapper[4744]: I1008 09:55:39.019868 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-utilities\") pod \"redhat-operators-4wsmr\" (UID: \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\") " pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:39 crc kubenswrapper[4744]: I1008 09:55:39.020120 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-utilities\") pod \"redhat-operators-4wsmr\" (UID: \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\") " pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:39 crc kubenswrapper[4744]: I1008 09:55:39.038922 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdchb\" (UniqueName: \"kubernetes.io/projected/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-kube-api-access-zdchb\") pod \"redhat-operators-4wsmr\" (UID: \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\") " pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:39 crc kubenswrapper[4744]: I1008 09:55:39.091307 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:39 crc kubenswrapper[4744]: I1008 09:55:39.524453 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4wsmr"] Oct 08 09:55:39 crc kubenswrapper[4744]: I1008 09:55:39.899587 4744 generic.go:334] "Generic (PLEG): container finished" podID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerID="5a97b330a6dd1beb74a1743ed77613338e849261d1d3f85a8821c209262bef68" exitCode=0 Oct 08 09:55:39 crc kubenswrapper[4744]: I1008 09:55:39.899904 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wsmr" event={"ID":"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12","Type":"ContainerDied","Data":"5a97b330a6dd1beb74a1743ed77613338e849261d1d3f85a8821c209262bef68"} Oct 08 09:55:39 crc kubenswrapper[4744]: I1008 09:55:39.899929 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wsmr" event={"ID":"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12","Type":"ContainerStarted","Data":"200eb3234b1eb6392fc808bcea43ca2759e3eac3824524a801e399c3045507d0"} Oct 08 09:55:41 crc kubenswrapper[4744]: I1008 09:55:41.927743 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wsmr" event={"ID":"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12","Type":"ContainerStarted","Data":"d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685"} Oct 08 09:55:44 crc kubenswrapper[4744]: E1008 09:55:44.593859 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2fbe2ab_1ea9_4d4b_a088_b267a9e07f12.slice/crio-d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2fbe2ab_1ea9_4d4b_a088_b267a9e07f12.slice/crio-conmon-d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685.scope\": RecentStats: unable to find data in memory cache]" Oct 08 09:55:44 crc kubenswrapper[4744]: I1008 09:55:44.961017 4744 generic.go:334] "Generic (PLEG): container finished" podID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerID="d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685" exitCode=0 Oct 08 09:55:44 crc kubenswrapper[4744]: I1008 09:55:44.961116 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wsmr" event={"ID":"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12","Type":"ContainerDied","Data":"d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685"} Oct 08 09:55:45 crc kubenswrapper[4744]: I1008 09:55:45.971634 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wsmr" event={"ID":"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12","Type":"ContainerStarted","Data":"a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5"} Oct 08 09:55:45 crc kubenswrapper[4744]: I1008 09:55:45.995032 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4wsmr" podStartSLOduration=2.516550239 podStartE2EDuration="7.995014076s" podCreationTimestamp="2025-10-08 09:55:38 +0000 UTC" firstStartedPulling="2025-10-08 09:55:39.906269868 +0000 UTC m=+2635.153915107" lastFinishedPulling="2025-10-08 09:55:45.384733705 +0000 UTC m=+2640.632378944" observedRunningTime="2025-10-08 09:55:45.988744092 +0000 UTC m=+2641.236389341" watchObservedRunningTime="2025-10-08 09:55:45.995014076 +0000 UTC m=+2641.242659315" Oct 08 09:55:49 crc kubenswrapper[4744]: I1008 09:55:49.091633 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:49 crc kubenswrapper[4744]: I1008 09:55:49.093105 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:55:50 crc kubenswrapper[4744]: I1008 09:55:50.138522 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4wsmr" podUID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerName="registry-server" probeResult="failure" output=< Oct 08 09:55:50 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 09:55:50 crc kubenswrapper[4744]: > Oct 08 09:55:56 crc kubenswrapper[4744]: I1008 09:55:56.059304 4744 generic.go:334] "Generic (PLEG): container finished" podID="3f2cc21c-f2e1-4775-8563-c794073928f4" containerID="a9a22fd10bf3b4dd058539065baae01a8000019115375d154c502b00ba325e17" exitCode=0 Oct 08 09:55:56 crc kubenswrapper[4744]: I1008 09:55:56.059362 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" event={"ID":"3f2cc21c-f2e1-4775-8563-c794073928f4","Type":"ContainerDied","Data":"a9a22fd10bf3b4dd058539065baae01a8000019115375d154c502b00ba325e17"} Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.464276 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.548614 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-neutron-metadata-combined-ca-bundle\") pod \"3f2cc21c-f2e1-4775-8563-c794073928f4\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.548919 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxwjl\" (UniqueName: \"kubernetes.io/projected/3f2cc21c-f2e1-4775-8563-c794073928f4-kube-api-access-zxwjl\") pod \"3f2cc21c-f2e1-4775-8563-c794073928f4\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.549060 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-inventory\") pod \"3f2cc21c-f2e1-4775-8563-c794073928f4\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.549200 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-neutron-ovn-metadata-agent-neutron-config-0\") pod \"3f2cc21c-f2e1-4775-8563-c794073928f4\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.549332 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-ssh-key\") pod \"3f2cc21c-f2e1-4775-8563-c794073928f4\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.549437 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-ceph\") pod \"3f2cc21c-f2e1-4775-8563-c794073928f4\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.549530 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-nova-metadata-neutron-config-0\") pod \"3f2cc21c-f2e1-4775-8563-c794073928f4\" (UID: \"3f2cc21c-f2e1-4775-8563-c794073928f4\") " Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.557831 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-ceph" (OuterVolumeSpecName: "ceph") pod "3f2cc21c-f2e1-4775-8563-c794073928f4" (UID: "3f2cc21c-f2e1-4775-8563-c794073928f4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.567552 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3f2cc21c-f2e1-4775-8563-c794073928f4" (UID: "3f2cc21c-f2e1-4775-8563-c794073928f4"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.567556 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f2cc21c-f2e1-4775-8563-c794073928f4-kube-api-access-zxwjl" (OuterVolumeSpecName: "kube-api-access-zxwjl") pod "3f2cc21c-f2e1-4775-8563-c794073928f4" (UID: "3f2cc21c-f2e1-4775-8563-c794073928f4"). InnerVolumeSpecName "kube-api-access-zxwjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.580325 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "3f2cc21c-f2e1-4775-8563-c794073928f4" (UID: "3f2cc21c-f2e1-4775-8563-c794073928f4"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.582353 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-inventory" (OuterVolumeSpecName: "inventory") pod "3f2cc21c-f2e1-4775-8563-c794073928f4" (UID: "3f2cc21c-f2e1-4775-8563-c794073928f4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.593865 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "3f2cc21c-f2e1-4775-8563-c794073928f4" (UID: "3f2cc21c-f2e1-4775-8563-c794073928f4"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.605647 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3f2cc21c-f2e1-4775-8563-c794073928f4" (UID: "3f2cc21c-f2e1-4775-8563-c794073928f4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.657851 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.657885 4744 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.657896 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.657905 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.657915 4744 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.657924 4744 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2cc21c-f2e1-4775-8563-c794073928f4-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 09:55:57 crc kubenswrapper[4744]: I1008 09:55:57.657933 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxwjl\" (UniqueName: \"kubernetes.io/projected/3f2cc21c-f2e1-4775-8563-c794073928f4-kube-api-access-zxwjl\") on node \"crc\" DevicePath \"\"" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.074303 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" event={"ID":"3f2cc21c-f2e1-4775-8563-c794073928f4","Type":"ContainerDied","Data":"d77535d18c078710fc418470480735261bc35c1a27db1ce890ddc05a2cabf223"} Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.074342 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d77535d18c078710fc418470480735261bc35c1a27db1ce890ddc05a2cabf223" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.074412 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.189043 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc"] Oct 08 09:55:58 crc kubenswrapper[4744]: E1008 09:55:58.189779 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2cc21c-f2e1-4775-8563-c794073928f4" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.189875 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2cc21c-f2e1-4775-8563-c794073928f4" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.190105 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2cc21c-f2e1-4775-8563-c794073928f4" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.190779 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.205060 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.205255 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.205491 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.206655 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.206692 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.206848 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.213564 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc"] Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.268644 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.268702 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.268730 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.268826 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skrxn\" (UniqueName: \"kubernetes.io/projected/e80d5289-7657-4835-8b43-936537129935-kube-api-access-skrxn\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.269030 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.269086 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.371338 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.371472 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skrxn\" (UniqueName: \"kubernetes.io/projected/e80d5289-7657-4835-8b43-936537129935-kube-api-access-skrxn\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.371580 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.371609 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.371694 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.371723 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.376680 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.377186 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.377439 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.380034 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.386506 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.392099 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skrxn\" (UniqueName: \"kubernetes.io/projected/e80d5289-7657-4835-8b43-936537129935-kube-api-access-skrxn\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-swsrc\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:58 crc kubenswrapper[4744]: I1008 09:55:58.504513 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 09:55:59 crc kubenswrapper[4744]: I1008 09:55:59.037494 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc"] Oct 08 09:55:59 crc kubenswrapper[4744]: I1008 09:55:59.090384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" event={"ID":"e80d5289-7657-4835-8b43-936537129935","Type":"ContainerStarted","Data":"8a2ce474765a945e448fba4234681c7df5e4566495749482dbb252eee954cd60"} Oct 08 09:56:00 crc kubenswrapper[4744]: I1008 09:56:00.100123 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" event={"ID":"e80d5289-7657-4835-8b43-936537129935","Type":"ContainerStarted","Data":"aecd8b16e43f68379cfc120d16e7eb8185be6c508570bc5111b6f53bef6973a8"} Oct 08 09:56:00 crc kubenswrapper[4744]: I1008 09:56:00.142706 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4wsmr" podUID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerName="registry-server" probeResult="failure" output=< Oct 08 09:56:00 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 09:56:00 crc kubenswrapper[4744]: > Oct 08 09:56:09 crc kubenswrapper[4744]: I1008 09:56:09.162261 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:56:09 crc kubenswrapper[4744]: I1008 09:56:09.196290 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" podStartSLOduration=10.669013177 podStartE2EDuration="11.196271598s" podCreationTimestamp="2025-10-08 09:55:58 +0000 UTC" firstStartedPulling="2025-10-08 09:55:59.051168832 +0000 UTC m=+2654.298814071" lastFinishedPulling="2025-10-08 09:55:59.578427253 +0000 UTC m=+2654.826072492" observedRunningTime="2025-10-08 09:56:00.120702891 +0000 UTC m=+2655.368348130" watchObservedRunningTime="2025-10-08 09:56:09.196271598 +0000 UTC m=+2664.443916837" Oct 08 09:56:09 crc kubenswrapper[4744]: I1008 09:56:09.222801 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:56:09 crc kubenswrapper[4744]: I1008 09:56:09.970551 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4wsmr"] Oct 08 09:56:10 crc kubenswrapper[4744]: I1008 09:56:10.188246 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4wsmr" podUID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerName="registry-server" containerID="cri-o://a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5" gracePeriod=2 Oct 08 09:56:10 crc kubenswrapper[4744]: I1008 09:56:10.653169 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:56:10 crc kubenswrapper[4744]: I1008 09:56:10.762267 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-utilities\") pod \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\" (UID: \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\") " Oct 08 09:56:10 crc kubenswrapper[4744]: I1008 09:56:10.762310 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-catalog-content\") pod \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\" (UID: \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\") " Oct 08 09:56:10 crc kubenswrapper[4744]: I1008 09:56:10.762337 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdchb\" (UniqueName: \"kubernetes.io/projected/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-kube-api-access-zdchb\") pod \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\" (UID: \"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12\") " Oct 08 09:56:10 crc kubenswrapper[4744]: I1008 09:56:10.763665 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-utilities" (OuterVolumeSpecName: "utilities") pod "f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" (UID: "f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:56:10 crc kubenswrapper[4744]: I1008 09:56:10.770644 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-kube-api-access-zdchb" (OuterVolumeSpecName: "kube-api-access-zdchb") pod "f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" (UID: "f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12"). InnerVolumeSpecName "kube-api-access-zdchb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:56:10 crc kubenswrapper[4744]: I1008 09:56:10.836480 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" (UID: "f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:56:10 crc kubenswrapper[4744]: I1008 09:56:10.864043 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:56:10 crc kubenswrapper[4744]: I1008 09:56:10.864091 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:56:10 crc kubenswrapper[4744]: I1008 09:56:10.864108 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdchb\" (UniqueName: \"kubernetes.io/projected/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12-kube-api-access-zdchb\") on node \"crc\" DevicePath \"\"" Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.201654 4744 generic.go:334] "Generic (PLEG): container finished" podID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerID="a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5" exitCode=0 Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.201696 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wsmr" event={"ID":"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12","Type":"ContainerDied","Data":"a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5"} Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.201721 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wsmr" event={"ID":"f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12","Type":"ContainerDied","Data":"200eb3234b1eb6392fc808bcea43ca2759e3eac3824524a801e399c3045507d0"} Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.201739 4744 scope.go:117] "RemoveContainer" containerID="a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5" Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.201892 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wsmr" Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.227007 4744 scope.go:117] "RemoveContainer" containerID="d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685" Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.275349 4744 scope.go:117] "RemoveContainer" containerID="5a97b330a6dd1beb74a1743ed77613338e849261d1d3f85a8821c209262bef68" Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.296226 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4wsmr"] Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.304866 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4wsmr"] Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.312327 4744 scope.go:117] "RemoveContainer" containerID="a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5" Oct 08 09:56:11 crc kubenswrapper[4744]: E1008 09:56:11.312949 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5\": container with ID starting with a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5 not found: ID does not exist" containerID="a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5" Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.312983 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5"} err="failed to get container status \"a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5\": rpc error: code = NotFound desc = could not find container \"a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5\": container with ID starting with a4ce71536a0921882e590c8eb1e76771a165768c25f9a7833309296c4aa24dc5 not found: ID does not exist" Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.313012 4744 scope.go:117] "RemoveContainer" containerID="d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685" Oct 08 09:56:11 crc kubenswrapper[4744]: E1008 09:56:11.313492 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685\": container with ID starting with d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685 not found: ID does not exist" containerID="d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685" Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.313565 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685"} err="failed to get container status \"d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685\": rpc error: code = NotFound desc = could not find container \"d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685\": container with ID starting with d6d5b7139cfd986a6106d743e673b1be52b22afd61d08feae9440c8f7e89d685 not found: ID does not exist" Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.313611 4744 scope.go:117] "RemoveContainer" containerID="5a97b330a6dd1beb74a1743ed77613338e849261d1d3f85a8821c209262bef68" Oct 08 09:56:11 crc kubenswrapper[4744]: E1008 09:56:11.314036 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a97b330a6dd1beb74a1743ed77613338e849261d1d3f85a8821c209262bef68\": container with ID starting with 5a97b330a6dd1beb74a1743ed77613338e849261d1d3f85a8821c209262bef68 not found: ID does not exist" containerID="5a97b330a6dd1beb74a1743ed77613338e849261d1d3f85a8821c209262bef68" Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.314059 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a97b330a6dd1beb74a1743ed77613338e849261d1d3f85a8821c209262bef68"} err="failed to get container status \"5a97b330a6dd1beb74a1743ed77613338e849261d1d3f85a8821c209262bef68\": rpc error: code = NotFound desc = could not find container \"5a97b330a6dd1beb74a1743ed77613338e849261d1d3f85a8821c209262bef68\": container with ID starting with 5a97b330a6dd1beb74a1743ed77613338e849261d1d3f85a8821c209262bef68 not found: ID does not exist" Oct 08 09:56:11 crc kubenswrapper[4744]: I1008 09:56:11.463633 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" path="/var/lib/kubelet/pods/f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12/volumes" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.128263 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xw9qr"] Oct 08 09:57:07 crc kubenswrapper[4744]: E1008 09:57:07.129350 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerName="registry-server" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.129366 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerName="registry-server" Oct 08 09:57:07 crc kubenswrapper[4744]: E1008 09:57:07.129407 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerName="extract-content" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.129415 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerName="extract-content" Oct 08 09:57:07 crc kubenswrapper[4744]: E1008 09:57:07.129430 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerName="extract-utilities" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.129437 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerName="extract-utilities" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.129664 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2fbe2ab-1ea9-4d4b-a088-b267a9e07f12" containerName="registry-server" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.131395 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.140079 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xw9qr"] Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.304949 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fl8w\" (UniqueName: \"kubernetes.io/projected/9cf41be3-4efa-4a18-b017-3f6000e7aabb-kube-api-access-5fl8w\") pod \"redhat-marketplace-xw9qr\" (UID: \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\") " pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.305038 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf41be3-4efa-4a18-b017-3f6000e7aabb-utilities\") pod \"redhat-marketplace-xw9qr\" (UID: \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\") " pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.305164 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf41be3-4efa-4a18-b017-3f6000e7aabb-catalog-content\") pod \"redhat-marketplace-xw9qr\" (UID: \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\") " pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.406795 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf41be3-4efa-4a18-b017-3f6000e7aabb-catalog-content\") pod \"redhat-marketplace-xw9qr\" (UID: \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\") " pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.406867 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fl8w\" (UniqueName: \"kubernetes.io/projected/9cf41be3-4efa-4a18-b017-3f6000e7aabb-kube-api-access-5fl8w\") pod \"redhat-marketplace-xw9qr\" (UID: \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\") " pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.406938 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf41be3-4efa-4a18-b017-3f6000e7aabb-utilities\") pod \"redhat-marketplace-xw9qr\" (UID: \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\") " pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.407747 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf41be3-4efa-4a18-b017-3f6000e7aabb-catalog-content\") pod \"redhat-marketplace-xw9qr\" (UID: \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\") " pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.407843 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf41be3-4efa-4a18-b017-3f6000e7aabb-utilities\") pod \"redhat-marketplace-xw9qr\" (UID: \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\") " pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.425275 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fl8w\" (UniqueName: \"kubernetes.io/projected/9cf41be3-4efa-4a18-b017-3f6000e7aabb-kube-api-access-5fl8w\") pod \"redhat-marketplace-xw9qr\" (UID: \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\") " pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.459571 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:07 crc kubenswrapper[4744]: I1008 09:57:07.957144 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xw9qr"] Oct 08 09:57:08 crc kubenswrapper[4744]: I1008 09:57:08.742573 4744 generic.go:334] "Generic (PLEG): container finished" podID="9cf41be3-4efa-4a18-b017-3f6000e7aabb" containerID="e3a4f10c59d0860bdb40b3ee1eeacfbe70f2d24cee62cd742e91c83cdc2161af" exitCode=0 Oct 08 09:57:08 crc kubenswrapper[4744]: I1008 09:57:08.742695 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xw9qr" event={"ID":"9cf41be3-4efa-4a18-b017-3f6000e7aabb","Type":"ContainerDied","Data":"e3a4f10c59d0860bdb40b3ee1eeacfbe70f2d24cee62cd742e91c83cdc2161af"} Oct 08 09:57:08 crc kubenswrapper[4744]: I1008 09:57:08.743061 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xw9qr" event={"ID":"9cf41be3-4efa-4a18-b017-3f6000e7aabb","Type":"ContainerStarted","Data":"64d47d70018bd63732461f142ec717a3a448d1ce58ece90d6729585ed35664d0"} Oct 08 09:57:09 crc kubenswrapper[4744]: I1008 09:57:09.756213 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xw9qr" event={"ID":"9cf41be3-4efa-4a18-b017-3f6000e7aabb","Type":"ContainerStarted","Data":"e7887a920e3dde6a07e5a57f84064f32e49ac308ad798712d7d5c4ebf513d3a5"} Oct 08 09:57:10 crc kubenswrapper[4744]: I1008 09:57:10.770349 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xw9qr" event={"ID":"9cf41be3-4efa-4a18-b017-3f6000e7aabb","Type":"ContainerDied","Data":"e7887a920e3dde6a07e5a57f84064f32e49ac308ad798712d7d5c4ebf513d3a5"} Oct 08 09:57:10 crc kubenswrapper[4744]: I1008 09:57:10.770690 4744 generic.go:334] "Generic (PLEG): container finished" podID="9cf41be3-4efa-4a18-b017-3f6000e7aabb" containerID="e7887a920e3dde6a07e5a57f84064f32e49ac308ad798712d7d5c4ebf513d3a5" exitCode=0 Oct 08 09:57:11 crc kubenswrapper[4744]: I1008 09:57:11.779743 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xw9qr" event={"ID":"9cf41be3-4efa-4a18-b017-3f6000e7aabb","Type":"ContainerStarted","Data":"f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd"} Oct 08 09:57:11 crc kubenswrapper[4744]: I1008 09:57:11.812963 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xw9qr" podStartSLOduration=2.263123679 podStartE2EDuration="4.812942427s" podCreationTimestamp="2025-10-08 09:57:07 +0000 UTC" firstStartedPulling="2025-10-08 09:57:08.74958481 +0000 UTC m=+2723.997230069" lastFinishedPulling="2025-10-08 09:57:11.299403578 +0000 UTC m=+2726.547048817" observedRunningTime="2025-10-08 09:57:11.806230281 +0000 UTC m=+2727.053875530" watchObservedRunningTime="2025-10-08 09:57:11.812942427 +0000 UTC m=+2727.060587666" Oct 08 09:57:17 crc kubenswrapper[4744]: I1008 09:57:17.467410 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:17 crc kubenswrapper[4744]: I1008 09:57:17.467961 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:17 crc kubenswrapper[4744]: I1008 09:57:17.536125 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:17 crc kubenswrapper[4744]: I1008 09:57:17.927307 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:17 crc kubenswrapper[4744]: I1008 09:57:17.976906 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xw9qr"] Oct 08 09:57:19 crc kubenswrapper[4744]: I1008 09:57:19.690074 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:57:19 crc kubenswrapper[4744]: I1008 09:57:19.691397 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:57:19 crc kubenswrapper[4744]: I1008 09:57:19.896848 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xw9qr" podUID="9cf41be3-4efa-4a18-b017-3f6000e7aabb" containerName="registry-server" containerID="cri-o://f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd" gracePeriod=2 Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.316861 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.475334 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fl8w\" (UniqueName: \"kubernetes.io/projected/9cf41be3-4efa-4a18-b017-3f6000e7aabb-kube-api-access-5fl8w\") pod \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\" (UID: \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\") " Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.475546 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf41be3-4efa-4a18-b017-3f6000e7aabb-utilities\") pod \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\" (UID: \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\") " Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.475592 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf41be3-4efa-4a18-b017-3f6000e7aabb-catalog-content\") pod \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\" (UID: \"9cf41be3-4efa-4a18-b017-3f6000e7aabb\") " Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.476429 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cf41be3-4efa-4a18-b017-3f6000e7aabb-utilities" (OuterVolumeSpecName: "utilities") pod "9cf41be3-4efa-4a18-b017-3f6000e7aabb" (UID: "9cf41be3-4efa-4a18-b017-3f6000e7aabb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.481968 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cf41be3-4efa-4a18-b017-3f6000e7aabb-kube-api-access-5fl8w" (OuterVolumeSpecName: "kube-api-access-5fl8w") pod "9cf41be3-4efa-4a18-b017-3f6000e7aabb" (UID: "9cf41be3-4efa-4a18-b017-3f6000e7aabb"). InnerVolumeSpecName "kube-api-access-5fl8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.489892 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cf41be3-4efa-4a18-b017-3f6000e7aabb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9cf41be3-4efa-4a18-b017-3f6000e7aabb" (UID: "9cf41be3-4efa-4a18-b017-3f6000e7aabb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.577811 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf41be3-4efa-4a18-b017-3f6000e7aabb-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.577846 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf41be3-4efa-4a18-b017-3f6000e7aabb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.577860 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fl8w\" (UniqueName: \"kubernetes.io/projected/9cf41be3-4efa-4a18-b017-3f6000e7aabb-kube-api-access-5fl8w\") on node \"crc\" DevicePath \"\"" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.909661 4744 generic.go:334] "Generic (PLEG): container finished" podID="9cf41be3-4efa-4a18-b017-3f6000e7aabb" containerID="f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd" exitCode=0 Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.909715 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xw9qr" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.909738 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xw9qr" event={"ID":"9cf41be3-4efa-4a18-b017-3f6000e7aabb","Type":"ContainerDied","Data":"f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd"} Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.910560 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xw9qr" event={"ID":"9cf41be3-4efa-4a18-b017-3f6000e7aabb","Type":"ContainerDied","Data":"64d47d70018bd63732461f142ec717a3a448d1ce58ece90d6729585ed35664d0"} Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.910580 4744 scope.go:117] "RemoveContainer" containerID="f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.928181 4744 scope.go:117] "RemoveContainer" containerID="e7887a920e3dde6a07e5a57f84064f32e49ac308ad798712d7d5c4ebf513d3a5" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.957572 4744 scope.go:117] "RemoveContainer" containerID="e3a4f10c59d0860bdb40b3ee1eeacfbe70f2d24cee62cd742e91c83cdc2161af" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.964738 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xw9qr"] Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.970529 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xw9qr"] Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.988525 4744 scope.go:117] "RemoveContainer" containerID="f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd" Oct 08 09:57:20 crc kubenswrapper[4744]: E1008 09:57:20.988949 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd\": container with ID starting with f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd not found: ID does not exist" containerID="f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.988997 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd"} err="failed to get container status \"f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd\": rpc error: code = NotFound desc = could not find container \"f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd\": container with ID starting with f236b30c4566013ba083a080a7df8ed46a4b2914cf45248ac42fc3fdb29fcdfd not found: ID does not exist" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.989026 4744 scope.go:117] "RemoveContainer" containerID="e7887a920e3dde6a07e5a57f84064f32e49ac308ad798712d7d5c4ebf513d3a5" Oct 08 09:57:20 crc kubenswrapper[4744]: E1008 09:57:20.989616 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7887a920e3dde6a07e5a57f84064f32e49ac308ad798712d7d5c4ebf513d3a5\": container with ID starting with e7887a920e3dde6a07e5a57f84064f32e49ac308ad798712d7d5c4ebf513d3a5 not found: ID does not exist" containerID="e7887a920e3dde6a07e5a57f84064f32e49ac308ad798712d7d5c4ebf513d3a5" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.989648 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7887a920e3dde6a07e5a57f84064f32e49ac308ad798712d7d5c4ebf513d3a5"} err="failed to get container status \"e7887a920e3dde6a07e5a57f84064f32e49ac308ad798712d7d5c4ebf513d3a5\": rpc error: code = NotFound desc = could not find container \"e7887a920e3dde6a07e5a57f84064f32e49ac308ad798712d7d5c4ebf513d3a5\": container with ID starting with e7887a920e3dde6a07e5a57f84064f32e49ac308ad798712d7d5c4ebf513d3a5 not found: ID does not exist" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.989669 4744 scope.go:117] "RemoveContainer" containerID="e3a4f10c59d0860bdb40b3ee1eeacfbe70f2d24cee62cd742e91c83cdc2161af" Oct 08 09:57:20 crc kubenswrapper[4744]: E1008 09:57:20.990065 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3a4f10c59d0860bdb40b3ee1eeacfbe70f2d24cee62cd742e91c83cdc2161af\": container with ID starting with e3a4f10c59d0860bdb40b3ee1eeacfbe70f2d24cee62cd742e91c83cdc2161af not found: ID does not exist" containerID="e3a4f10c59d0860bdb40b3ee1eeacfbe70f2d24cee62cd742e91c83cdc2161af" Oct 08 09:57:20 crc kubenswrapper[4744]: I1008 09:57:20.990223 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3a4f10c59d0860bdb40b3ee1eeacfbe70f2d24cee62cd742e91c83cdc2161af"} err="failed to get container status \"e3a4f10c59d0860bdb40b3ee1eeacfbe70f2d24cee62cd742e91c83cdc2161af\": rpc error: code = NotFound desc = could not find container \"e3a4f10c59d0860bdb40b3ee1eeacfbe70f2d24cee62cd742e91c83cdc2161af\": container with ID starting with e3a4f10c59d0860bdb40b3ee1eeacfbe70f2d24cee62cd742e91c83cdc2161af not found: ID does not exist" Oct 08 09:57:21 crc kubenswrapper[4744]: I1008 09:57:21.464749 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cf41be3-4efa-4a18-b017-3f6000e7aabb" path="/var/lib/kubelet/pods/9cf41be3-4efa-4a18-b017-3f6000e7aabb/volumes" Oct 08 09:57:49 crc kubenswrapper[4744]: I1008 09:57:49.690316 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:57:49 crc kubenswrapper[4744]: I1008 09:57:49.690985 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:58:19 crc kubenswrapper[4744]: I1008 09:58:19.690750 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 09:58:19 crc kubenswrapper[4744]: I1008 09:58:19.691306 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 09:58:19 crc kubenswrapper[4744]: I1008 09:58:19.691352 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 09:58:19 crc kubenswrapper[4744]: I1008 09:58:19.691818 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa495b33f33b23e8dcca10511302304051a76ce3efd1ef1a213237865410ba57"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 09:58:19 crc kubenswrapper[4744]: I1008 09:58:19.691872 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://aa495b33f33b23e8dcca10511302304051a76ce3efd1ef1a213237865410ba57" gracePeriod=600 Oct 08 09:58:20 crc kubenswrapper[4744]: I1008 09:58:20.440904 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="aa495b33f33b23e8dcca10511302304051a76ce3efd1ef1a213237865410ba57" exitCode=0 Oct 08 09:58:20 crc kubenswrapper[4744]: I1008 09:58:20.440966 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"aa495b33f33b23e8dcca10511302304051a76ce3efd1ef1a213237865410ba57"} Oct 08 09:58:20 crc kubenswrapper[4744]: I1008 09:58:20.441721 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8"} Oct 08 09:58:20 crc kubenswrapper[4744]: I1008 09:58:20.441758 4744 scope.go:117] "RemoveContainer" containerID="43b4780a857f3b61b804e0c937120336a12ad077902b0cfee51765f8abfc18af" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.160298 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4"] Oct 08 10:00:00 crc kubenswrapper[4744]: E1008 10:00:00.162094 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf41be3-4efa-4a18-b017-3f6000e7aabb" containerName="registry-server" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.162126 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf41be3-4efa-4a18-b017-3f6000e7aabb" containerName="registry-server" Oct 08 10:00:00 crc kubenswrapper[4744]: E1008 10:00:00.162155 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf41be3-4efa-4a18-b017-3f6000e7aabb" containerName="extract-utilities" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.162167 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf41be3-4efa-4a18-b017-3f6000e7aabb" containerName="extract-utilities" Oct 08 10:00:00 crc kubenswrapper[4744]: E1008 10:00:00.162189 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf41be3-4efa-4a18-b017-3f6000e7aabb" containerName="extract-content" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.162197 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf41be3-4efa-4a18-b017-3f6000e7aabb" containerName="extract-content" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.162534 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cf41be3-4efa-4a18-b017-3f6000e7aabb" containerName="registry-server" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.163723 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.167865 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.168364 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.182233 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4"] Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.358456 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-secret-volume\") pod \"collect-profiles-29331960-mcht4\" (UID: \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.358526 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-config-volume\") pod \"collect-profiles-29331960-mcht4\" (UID: \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.358552 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rssj7\" (UniqueName: \"kubernetes.io/projected/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-kube-api-access-rssj7\") pod \"collect-profiles-29331960-mcht4\" (UID: \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.461278 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-secret-volume\") pod \"collect-profiles-29331960-mcht4\" (UID: \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.461435 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-config-volume\") pod \"collect-profiles-29331960-mcht4\" (UID: \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.461470 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rssj7\" (UniqueName: \"kubernetes.io/projected/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-kube-api-access-rssj7\") pod \"collect-profiles-29331960-mcht4\" (UID: \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.463342 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-config-volume\") pod \"collect-profiles-29331960-mcht4\" (UID: \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.476675 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-secret-volume\") pod \"collect-profiles-29331960-mcht4\" (UID: \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.484679 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rssj7\" (UniqueName: \"kubernetes.io/projected/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-kube-api-access-rssj7\") pod \"collect-profiles-29331960-mcht4\" (UID: \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:00 crc kubenswrapper[4744]: I1008 10:00:00.487955 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:01 crc kubenswrapper[4744]: I1008 10:00:01.023331 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4"] Oct 08 10:00:01 crc kubenswrapper[4744]: I1008 10:00:01.367103 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" event={"ID":"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576","Type":"ContainerStarted","Data":"1d6454f7ea88c33aa0c91e67f30d2c1cb33fc0ed79dd6135cfb0b83d41726065"} Oct 08 10:00:01 crc kubenswrapper[4744]: I1008 10:00:01.367469 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" event={"ID":"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576","Type":"ContainerStarted","Data":"4ba2d5d3fa2beb7426c080bdb9d4a923df808c4359d60d9e92a731241a9f9f9f"} Oct 08 10:00:01 crc kubenswrapper[4744]: I1008 10:00:01.392996 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" podStartSLOduration=1.392974922 podStartE2EDuration="1.392974922s" podCreationTimestamp="2025-10-08 10:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:00:01.389734162 +0000 UTC m=+2896.637379401" watchObservedRunningTime="2025-10-08 10:00:01.392974922 +0000 UTC m=+2896.640620161" Oct 08 10:00:02 crc kubenswrapper[4744]: I1008 10:00:02.377616 4744 generic.go:334] "Generic (PLEG): container finished" podID="aa6ea4af-8627-4dde-ae7b-bb8ed25b1576" containerID="1d6454f7ea88c33aa0c91e67f30d2c1cb33fc0ed79dd6135cfb0b83d41726065" exitCode=0 Oct 08 10:00:02 crc kubenswrapper[4744]: I1008 10:00:02.377909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" event={"ID":"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576","Type":"ContainerDied","Data":"1d6454f7ea88c33aa0c91e67f30d2c1cb33fc0ed79dd6135cfb0b83d41726065"} Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.134888 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ffc48"] Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.136548 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.154293 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ffc48"] Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.211817 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b2c49b4-9d9a-421a-8e26-644c75f84777-utilities\") pod \"certified-operators-ffc48\" (UID: \"2b2c49b4-9d9a-421a-8e26-644c75f84777\") " pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.211958 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b2c49b4-9d9a-421a-8e26-644c75f84777-catalog-content\") pod \"certified-operators-ffc48\" (UID: \"2b2c49b4-9d9a-421a-8e26-644c75f84777\") " pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.212009 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv79z\" (UniqueName: \"kubernetes.io/projected/2b2c49b4-9d9a-421a-8e26-644c75f84777-kube-api-access-rv79z\") pod \"certified-operators-ffc48\" (UID: \"2b2c49b4-9d9a-421a-8e26-644c75f84777\") " pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.313960 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b2c49b4-9d9a-421a-8e26-644c75f84777-catalog-content\") pod \"certified-operators-ffc48\" (UID: \"2b2c49b4-9d9a-421a-8e26-644c75f84777\") " pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.314018 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv79z\" (UniqueName: \"kubernetes.io/projected/2b2c49b4-9d9a-421a-8e26-644c75f84777-kube-api-access-rv79z\") pod \"certified-operators-ffc48\" (UID: \"2b2c49b4-9d9a-421a-8e26-644c75f84777\") " pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.314125 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b2c49b4-9d9a-421a-8e26-644c75f84777-utilities\") pod \"certified-operators-ffc48\" (UID: \"2b2c49b4-9d9a-421a-8e26-644c75f84777\") " pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.314717 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b2c49b4-9d9a-421a-8e26-644c75f84777-catalog-content\") pod \"certified-operators-ffc48\" (UID: \"2b2c49b4-9d9a-421a-8e26-644c75f84777\") " pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.314780 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b2c49b4-9d9a-421a-8e26-644c75f84777-utilities\") pod \"certified-operators-ffc48\" (UID: \"2b2c49b4-9d9a-421a-8e26-644c75f84777\") " pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.344743 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv79z\" (UniqueName: \"kubernetes.io/projected/2b2c49b4-9d9a-421a-8e26-644c75f84777-kube-api-access-rv79z\") pod \"certified-operators-ffc48\" (UID: \"2b2c49b4-9d9a-421a-8e26-644c75f84777\") " pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.465294 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.716759 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.825011 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rssj7\" (UniqueName: \"kubernetes.io/projected/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-kube-api-access-rssj7\") pod \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\" (UID: \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\") " Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.825413 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-config-volume\") pod \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\" (UID: \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\") " Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.825537 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-secret-volume\") pod \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\" (UID: \"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576\") " Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.828216 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-config-volume" (OuterVolumeSpecName: "config-volume") pod "aa6ea4af-8627-4dde-ae7b-bb8ed25b1576" (UID: "aa6ea4af-8627-4dde-ae7b-bb8ed25b1576"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.840597 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "aa6ea4af-8627-4dde-ae7b-bb8ed25b1576" (UID: "aa6ea4af-8627-4dde-ae7b-bb8ed25b1576"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.843554 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-kube-api-access-rssj7" (OuterVolumeSpecName: "kube-api-access-rssj7") pod "aa6ea4af-8627-4dde-ae7b-bb8ed25b1576" (UID: "aa6ea4af-8627-4dde-ae7b-bb8ed25b1576"). InnerVolumeSpecName "kube-api-access-rssj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.928180 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rssj7\" (UniqueName: \"kubernetes.io/projected/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-kube-api-access-rssj7\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.928209 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:03 crc kubenswrapper[4744]: I1008 10:00:03.928220 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa6ea4af-8627-4dde-ae7b-bb8ed25b1576-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:04 crc kubenswrapper[4744]: I1008 10:00:04.099520 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ffc48"] Oct 08 10:00:04 crc kubenswrapper[4744]: W1008 10:00:04.104474 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b2c49b4_9d9a_421a_8e26_644c75f84777.slice/crio-a1c8b0ca29b3e8330d64c52a36aa197fc60c3e57a555d43c168542638a2a6794 WatchSource:0}: Error finding container a1c8b0ca29b3e8330d64c52a36aa197fc60c3e57a555d43c168542638a2a6794: Status 404 returned error can't find the container with id a1c8b0ca29b3e8330d64c52a36aa197fc60c3e57a555d43c168542638a2a6794 Oct 08 10:00:04 crc kubenswrapper[4744]: I1008 10:00:04.393856 4744 generic.go:334] "Generic (PLEG): container finished" podID="2b2c49b4-9d9a-421a-8e26-644c75f84777" containerID="3a2c0dd3dc216514e400edafa4ea1b376b72146cf4f47ef236b0cfd3ca1311d0" exitCode=0 Oct 08 10:00:04 crc kubenswrapper[4744]: I1008 10:00:04.393920 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffc48" event={"ID":"2b2c49b4-9d9a-421a-8e26-644c75f84777","Type":"ContainerDied","Data":"3a2c0dd3dc216514e400edafa4ea1b376b72146cf4f47ef236b0cfd3ca1311d0"} Oct 08 10:00:04 crc kubenswrapper[4744]: I1008 10:00:04.394214 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffc48" event={"ID":"2b2c49b4-9d9a-421a-8e26-644c75f84777","Type":"ContainerStarted","Data":"a1c8b0ca29b3e8330d64c52a36aa197fc60c3e57a555d43c168542638a2a6794"} Oct 08 10:00:04 crc kubenswrapper[4744]: I1008 10:00:04.397092 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 10:00:04 crc kubenswrapper[4744]: I1008 10:00:04.399634 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" event={"ID":"aa6ea4af-8627-4dde-ae7b-bb8ed25b1576","Type":"ContainerDied","Data":"4ba2d5d3fa2beb7426c080bdb9d4a923df808c4359d60d9e92a731241a9f9f9f"} Oct 08 10:00:04 crc kubenswrapper[4744]: I1008 10:00:04.399682 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ba2d5d3fa2beb7426c080bdb9d4a923df808c4359d60d9e92a731241a9f9f9f" Oct 08 10:00:04 crc kubenswrapper[4744]: I1008 10:00:04.399746 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331960-mcht4" Oct 08 10:00:04 crc kubenswrapper[4744]: I1008 10:00:04.508328 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2"] Oct 08 10:00:04 crc kubenswrapper[4744]: I1008 10:00:04.516129 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331915-7gnz2"] Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.411044 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffc48" event={"ID":"2b2c49b4-9d9a-421a-8e26-644c75f84777","Type":"ContainerStarted","Data":"5f411a2db74a9c3ce8f636d38d8bd734fa148af8fc6e6fca9451ee58e142039c"} Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.466152 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c847191-fb16-4cdc-8b54-ee09e3571cce" path="/var/lib/kubelet/pods/9c847191-fb16-4cdc-8b54-ee09e3571cce/volumes" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.562683 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8mnq9"] Oct 08 10:00:05 crc kubenswrapper[4744]: E1008 10:00:05.563110 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6ea4af-8627-4dde-ae7b-bb8ed25b1576" containerName="collect-profiles" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.563125 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6ea4af-8627-4dde-ae7b-bb8ed25b1576" containerName="collect-profiles" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.563292 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa6ea4af-8627-4dde-ae7b-bb8ed25b1576" containerName="collect-profiles" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.564622 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.586051 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8mnq9"] Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.765587 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-catalog-content\") pod \"community-operators-8mnq9\" (UID: \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\") " pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.765734 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-utilities\") pod \"community-operators-8mnq9\" (UID: \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\") " pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.766156 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcdnf\" (UniqueName: \"kubernetes.io/projected/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-kube-api-access-hcdnf\") pod \"community-operators-8mnq9\" (UID: \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\") " pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.868525 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-utilities\") pod \"community-operators-8mnq9\" (UID: \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\") " pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.868622 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcdnf\" (UniqueName: \"kubernetes.io/projected/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-kube-api-access-hcdnf\") pod \"community-operators-8mnq9\" (UID: \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\") " pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.868684 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-catalog-content\") pod \"community-operators-8mnq9\" (UID: \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\") " pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.869114 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-utilities\") pod \"community-operators-8mnq9\" (UID: \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\") " pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.869161 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-catalog-content\") pod \"community-operators-8mnq9\" (UID: \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\") " pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:05 crc kubenswrapper[4744]: I1008 10:00:05.907265 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcdnf\" (UniqueName: \"kubernetes.io/projected/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-kube-api-access-hcdnf\") pod \"community-operators-8mnq9\" (UID: \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\") " pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:06 crc kubenswrapper[4744]: I1008 10:00:06.187354 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:06 crc kubenswrapper[4744]: I1008 10:00:06.708389 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8mnq9"] Oct 08 10:00:06 crc kubenswrapper[4744]: W1008 10:00:06.712667 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd08d26cc_7c0a_45d5_bbf2_70ad2405a6f6.slice/crio-39be2827baaf910c9a386da09ecbd6cb6522ce619346dfba0d2f14cf78675cd8 WatchSource:0}: Error finding container 39be2827baaf910c9a386da09ecbd6cb6522ce619346dfba0d2f14cf78675cd8: Status 404 returned error can't find the container with id 39be2827baaf910c9a386da09ecbd6cb6522ce619346dfba0d2f14cf78675cd8 Oct 08 10:00:07 crc kubenswrapper[4744]: I1008 10:00:07.435278 4744 generic.go:334] "Generic (PLEG): container finished" podID="d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" containerID="e17f7c9d8012b42c2053f81b7e6a4c97d6e25594c17aefe4616c72aded327c8b" exitCode=0 Oct 08 10:00:07 crc kubenswrapper[4744]: I1008 10:00:07.435359 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mnq9" event={"ID":"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6","Type":"ContainerDied","Data":"e17f7c9d8012b42c2053f81b7e6a4c97d6e25594c17aefe4616c72aded327c8b"} Oct 08 10:00:07 crc kubenswrapper[4744]: I1008 10:00:07.435982 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mnq9" event={"ID":"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6","Type":"ContainerStarted","Data":"39be2827baaf910c9a386da09ecbd6cb6522ce619346dfba0d2f14cf78675cd8"} Oct 08 10:00:07 crc kubenswrapper[4744]: I1008 10:00:07.439330 4744 generic.go:334] "Generic (PLEG): container finished" podID="2b2c49b4-9d9a-421a-8e26-644c75f84777" containerID="5f411a2db74a9c3ce8f636d38d8bd734fa148af8fc6e6fca9451ee58e142039c" exitCode=0 Oct 08 10:00:07 crc kubenswrapper[4744]: I1008 10:00:07.439442 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffc48" event={"ID":"2b2c49b4-9d9a-421a-8e26-644c75f84777","Type":"ContainerDied","Data":"5f411a2db74a9c3ce8f636d38d8bd734fa148af8fc6e6fca9451ee58e142039c"} Oct 08 10:00:08 crc kubenswrapper[4744]: I1008 10:00:08.450699 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mnq9" event={"ID":"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6","Type":"ContainerStarted","Data":"4a1844e23cdf7093127eb072f0d78492ed28cf068138ac47eca67a66ffee84c4"} Oct 08 10:00:08 crc kubenswrapper[4744]: I1008 10:00:08.457136 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffc48" event={"ID":"2b2c49b4-9d9a-421a-8e26-644c75f84777","Type":"ContainerStarted","Data":"08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34"} Oct 08 10:00:08 crc kubenswrapper[4744]: I1008 10:00:08.492470 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ffc48" podStartSLOduration=1.986284987 podStartE2EDuration="5.492447465s" podCreationTimestamp="2025-10-08 10:00:03 +0000 UTC" firstStartedPulling="2025-10-08 10:00:04.396818144 +0000 UTC m=+2899.644463393" lastFinishedPulling="2025-10-08 10:00:07.902980642 +0000 UTC m=+2903.150625871" observedRunningTime="2025-10-08 10:00:08.489267926 +0000 UTC m=+2903.736913175" watchObservedRunningTime="2025-10-08 10:00:08.492447465 +0000 UTC m=+2903.740092704" Oct 08 10:00:10 crc kubenswrapper[4744]: I1008 10:00:10.481968 4744 generic.go:334] "Generic (PLEG): container finished" podID="d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" containerID="4a1844e23cdf7093127eb072f0d78492ed28cf068138ac47eca67a66ffee84c4" exitCode=0 Oct 08 10:00:10 crc kubenswrapper[4744]: I1008 10:00:10.482034 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mnq9" event={"ID":"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6","Type":"ContainerDied","Data":"4a1844e23cdf7093127eb072f0d78492ed28cf068138ac47eca67a66ffee84c4"} Oct 08 10:00:12 crc kubenswrapper[4744]: I1008 10:00:12.501798 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mnq9" event={"ID":"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6","Type":"ContainerStarted","Data":"32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57"} Oct 08 10:00:12 crc kubenswrapper[4744]: I1008 10:00:12.534394 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8mnq9" podStartSLOduration=3.540250653 podStartE2EDuration="7.534361051s" podCreationTimestamp="2025-10-08 10:00:05 +0000 UTC" firstStartedPulling="2025-10-08 10:00:07.43822921 +0000 UTC m=+2902.685874449" lastFinishedPulling="2025-10-08 10:00:11.432339608 +0000 UTC m=+2906.679984847" observedRunningTime="2025-10-08 10:00:12.527655075 +0000 UTC m=+2907.775300334" watchObservedRunningTime="2025-10-08 10:00:12.534361051 +0000 UTC m=+2907.782006290" Oct 08 10:00:13 crc kubenswrapper[4744]: I1008 10:00:13.466128 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:13 crc kubenswrapper[4744]: I1008 10:00:13.466486 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:13 crc kubenswrapper[4744]: I1008 10:00:13.529962 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:14 crc kubenswrapper[4744]: I1008 10:00:14.567407 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:14 crc kubenswrapper[4744]: I1008 10:00:14.936106 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ffc48"] Oct 08 10:00:16 crc kubenswrapper[4744]: I1008 10:00:16.188388 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:16 crc kubenswrapper[4744]: I1008 10:00:16.188471 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:16 crc kubenswrapper[4744]: I1008 10:00:16.240175 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:16 crc kubenswrapper[4744]: I1008 10:00:16.537246 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ffc48" podUID="2b2c49b4-9d9a-421a-8e26-644c75f84777" containerName="registry-server" containerID="cri-o://08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34" gracePeriod=2 Oct 08 10:00:16 crc kubenswrapper[4744]: I1008 10:00:16.597568 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.013584 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.097248 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b2c49b4-9d9a-421a-8e26-644c75f84777-catalog-content\") pod \"2b2c49b4-9d9a-421a-8e26-644c75f84777\" (UID: \"2b2c49b4-9d9a-421a-8e26-644c75f84777\") " Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.097833 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv79z\" (UniqueName: \"kubernetes.io/projected/2b2c49b4-9d9a-421a-8e26-644c75f84777-kube-api-access-rv79z\") pod \"2b2c49b4-9d9a-421a-8e26-644c75f84777\" (UID: \"2b2c49b4-9d9a-421a-8e26-644c75f84777\") " Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.097955 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b2c49b4-9d9a-421a-8e26-644c75f84777-utilities\") pod \"2b2c49b4-9d9a-421a-8e26-644c75f84777\" (UID: \"2b2c49b4-9d9a-421a-8e26-644c75f84777\") " Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.098838 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b2c49b4-9d9a-421a-8e26-644c75f84777-utilities" (OuterVolumeSpecName: "utilities") pod "2b2c49b4-9d9a-421a-8e26-644c75f84777" (UID: "2b2c49b4-9d9a-421a-8e26-644c75f84777"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.099877 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b2c49b4-9d9a-421a-8e26-644c75f84777-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.107758 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b2c49b4-9d9a-421a-8e26-644c75f84777-kube-api-access-rv79z" (OuterVolumeSpecName: "kube-api-access-rv79z") pod "2b2c49b4-9d9a-421a-8e26-644c75f84777" (UID: "2b2c49b4-9d9a-421a-8e26-644c75f84777"). InnerVolumeSpecName "kube-api-access-rv79z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.163856 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b2c49b4-9d9a-421a-8e26-644c75f84777-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b2c49b4-9d9a-421a-8e26-644c75f84777" (UID: "2b2c49b4-9d9a-421a-8e26-644c75f84777"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.205446 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b2c49b4-9d9a-421a-8e26-644c75f84777-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.205496 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv79z\" (UniqueName: \"kubernetes.io/projected/2b2c49b4-9d9a-421a-8e26-644c75f84777-kube-api-access-rv79z\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.550274 4744 generic.go:334] "Generic (PLEG): container finished" podID="2b2c49b4-9d9a-421a-8e26-644c75f84777" containerID="08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34" exitCode=0 Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.550342 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ffc48" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.550341 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffc48" event={"ID":"2b2c49b4-9d9a-421a-8e26-644c75f84777","Type":"ContainerDied","Data":"08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34"} Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.550911 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ffc48" event={"ID":"2b2c49b4-9d9a-421a-8e26-644c75f84777","Type":"ContainerDied","Data":"a1c8b0ca29b3e8330d64c52a36aa197fc60c3e57a555d43c168542638a2a6794"} Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.550943 4744 scope.go:117] "RemoveContainer" containerID="08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.578731 4744 scope.go:117] "RemoveContainer" containerID="5f411a2db74a9c3ce8f636d38d8bd734fa148af8fc6e6fca9451ee58e142039c" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.583403 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ffc48"] Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.591257 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ffc48"] Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.624417 4744 scope.go:117] "RemoveContainer" containerID="3a2c0dd3dc216514e400edafa4ea1b376b72146cf4f47ef236b0cfd3ca1311d0" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.651461 4744 scope.go:117] "RemoveContainer" containerID="08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34" Oct 08 10:00:17 crc kubenswrapper[4744]: E1008 10:00:17.651987 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34\": container with ID starting with 08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34 not found: ID does not exist" containerID="08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.652027 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34"} err="failed to get container status \"08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34\": rpc error: code = NotFound desc = could not find container \"08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34\": container with ID starting with 08be5c06ca8322f9a8862f3a780434d4c07dfbe4c4694dd04b626fe0b7cf0e34 not found: ID does not exist" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.652053 4744 scope.go:117] "RemoveContainer" containerID="5f411a2db74a9c3ce8f636d38d8bd734fa148af8fc6e6fca9451ee58e142039c" Oct 08 10:00:17 crc kubenswrapper[4744]: E1008 10:00:17.652472 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f411a2db74a9c3ce8f636d38d8bd734fa148af8fc6e6fca9451ee58e142039c\": container with ID starting with 5f411a2db74a9c3ce8f636d38d8bd734fa148af8fc6e6fca9451ee58e142039c not found: ID does not exist" containerID="5f411a2db74a9c3ce8f636d38d8bd734fa148af8fc6e6fca9451ee58e142039c" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.652502 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f411a2db74a9c3ce8f636d38d8bd734fa148af8fc6e6fca9451ee58e142039c"} err="failed to get container status \"5f411a2db74a9c3ce8f636d38d8bd734fa148af8fc6e6fca9451ee58e142039c\": rpc error: code = NotFound desc = could not find container \"5f411a2db74a9c3ce8f636d38d8bd734fa148af8fc6e6fca9451ee58e142039c\": container with ID starting with 5f411a2db74a9c3ce8f636d38d8bd734fa148af8fc6e6fca9451ee58e142039c not found: ID does not exist" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.652524 4744 scope.go:117] "RemoveContainer" containerID="3a2c0dd3dc216514e400edafa4ea1b376b72146cf4f47ef236b0cfd3ca1311d0" Oct 08 10:00:17 crc kubenswrapper[4744]: E1008 10:00:17.652788 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a2c0dd3dc216514e400edafa4ea1b376b72146cf4f47ef236b0cfd3ca1311d0\": container with ID starting with 3a2c0dd3dc216514e400edafa4ea1b376b72146cf4f47ef236b0cfd3ca1311d0 not found: ID does not exist" containerID="3a2c0dd3dc216514e400edafa4ea1b376b72146cf4f47ef236b0cfd3ca1311d0" Oct 08 10:00:17 crc kubenswrapper[4744]: I1008 10:00:17.652812 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a2c0dd3dc216514e400edafa4ea1b376b72146cf4f47ef236b0cfd3ca1311d0"} err="failed to get container status \"3a2c0dd3dc216514e400edafa4ea1b376b72146cf4f47ef236b0cfd3ca1311d0\": rpc error: code = NotFound desc = could not find container \"3a2c0dd3dc216514e400edafa4ea1b376b72146cf4f47ef236b0cfd3ca1311d0\": container with ID starting with 3a2c0dd3dc216514e400edafa4ea1b376b72146cf4f47ef236b0cfd3ca1311d0 not found: ID does not exist" Oct 08 10:00:18 crc kubenswrapper[4744]: I1008 10:00:18.513751 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8mnq9"] Oct 08 10:00:18 crc kubenswrapper[4744]: I1008 10:00:18.575548 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8mnq9" podUID="d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" containerName="registry-server" containerID="cri-o://32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57" gracePeriod=2 Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.005479 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.137760 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-utilities\") pod \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\" (UID: \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\") " Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.137845 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-catalog-content\") pod \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\" (UID: \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\") " Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.137970 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcdnf\" (UniqueName: \"kubernetes.io/projected/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-kube-api-access-hcdnf\") pod \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\" (UID: \"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6\") " Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.138699 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-utilities" (OuterVolumeSpecName: "utilities") pod "d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" (UID: "d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.147553 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-kube-api-access-hcdnf" (OuterVolumeSpecName: "kube-api-access-hcdnf") pod "d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" (UID: "d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6"). InnerVolumeSpecName "kube-api-access-hcdnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.195773 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" (UID: "d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.240114 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcdnf\" (UniqueName: \"kubernetes.io/projected/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-kube-api-access-hcdnf\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.240160 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.240170 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.463761 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b2c49b4-9d9a-421a-8e26-644c75f84777" path="/var/lib/kubelet/pods/2b2c49b4-9d9a-421a-8e26-644c75f84777/volumes" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.585047 4744 generic.go:334] "Generic (PLEG): container finished" podID="d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" containerID="32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57" exitCode=0 Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.585095 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mnq9" event={"ID":"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6","Type":"ContainerDied","Data":"32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57"} Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.585123 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8mnq9" event={"ID":"d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6","Type":"ContainerDied","Data":"39be2827baaf910c9a386da09ecbd6cb6522ce619346dfba0d2f14cf78675cd8"} Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.585142 4744 scope.go:117] "RemoveContainer" containerID="32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.585282 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8mnq9" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.643748 4744 scope.go:117] "RemoveContainer" containerID="4a1844e23cdf7093127eb072f0d78492ed28cf068138ac47eca67a66ffee84c4" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.649184 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8mnq9"] Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.657541 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8mnq9"] Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.670356 4744 scope.go:117] "RemoveContainer" containerID="e17f7c9d8012b42c2053f81b7e6a4c97d6e25594c17aefe4616c72aded327c8b" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.709151 4744 scope.go:117] "RemoveContainer" containerID="32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57" Oct 08 10:00:19 crc kubenswrapper[4744]: E1008 10:00:19.709552 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57\": container with ID starting with 32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57 not found: ID does not exist" containerID="32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.709604 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57"} err="failed to get container status \"32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57\": rpc error: code = NotFound desc = could not find container \"32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57\": container with ID starting with 32346f27942e102578df3b3aa4fe487822d08eca7b8c0299b3252837325a5d57 not found: ID does not exist" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.709637 4744 scope.go:117] "RemoveContainer" containerID="4a1844e23cdf7093127eb072f0d78492ed28cf068138ac47eca67a66ffee84c4" Oct 08 10:00:19 crc kubenswrapper[4744]: E1008 10:00:19.710032 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a1844e23cdf7093127eb072f0d78492ed28cf068138ac47eca67a66ffee84c4\": container with ID starting with 4a1844e23cdf7093127eb072f0d78492ed28cf068138ac47eca67a66ffee84c4 not found: ID does not exist" containerID="4a1844e23cdf7093127eb072f0d78492ed28cf068138ac47eca67a66ffee84c4" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.710062 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a1844e23cdf7093127eb072f0d78492ed28cf068138ac47eca67a66ffee84c4"} err="failed to get container status \"4a1844e23cdf7093127eb072f0d78492ed28cf068138ac47eca67a66ffee84c4\": rpc error: code = NotFound desc = could not find container \"4a1844e23cdf7093127eb072f0d78492ed28cf068138ac47eca67a66ffee84c4\": container with ID starting with 4a1844e23cdf7093127eb072f0d78492ed28cf068138ac47eca67a66ffee84c4 not found: ID does not exist" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.710081 4744 scope.go:117] "RemoveContainer" containerID="e17f7c9d8012b42c2053f81b7e6a4c97d6e25594c17aefe4616c72aded327c8b" Oct 08 10:00:19 crc kubenswrapper[4744]: E1008 10:00:19.710476 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e17f7c9d8012b42c2053f81b7e6a4c97d6e25594c17aefe4616c72aded327c8b\": container with ID starting with e17f7c9d8012b42c2053f81b7e6a4c97d6e25594c17aefe4616c72aded327c8b not found: ID does not exist" containerID="e17f7c9d8012b42c2053f81b7e6a4c97d6e25594c17aefe4616c72aded327c8b" Oct 08 10:00:19 crc kubenswrapper[4744]: I1008 10:00:19.710514 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e17f7c9d8012b42c2053f81b7e6a4c97d6e25594c17aefe4616c72aded327c8b"} err="failed to get container status \"e17f7c9d8012b42c2053f81b7e6a4c97d6e25594c17aefe4616c72aded327c8b\": rpc error: code = NotFound desc = could not find container \"e17f7c9d8012b42c2053f81b7e6a4c97d6e25594c17aefe4616c72aded327c8b\": container with ID starting with e17f7c9d8012b42c2053f81b7e6a4c97d6e25594c17aefe4616c72aded327c8b not found: ID does not exist" Oct 08 10:00:21 crc kubenswrapper[4744]: I1008 10:00:21.463428 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" path="/var/lib/kubelet/pods/d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6/volumes" Oct 08 10:00:35 crc kubenswrapper[4744]: I1008 10:00:35.713905 4744 generic.go:334] "Generic (PLEG): container finished" podID="e80d5289-7657-4835-8b43-936537129935" containerID="aecd8b16e43f68379cfc120d16e7eb8185be6c508570bc5111b6f53bef6973a8" exitCode=0 Oct 08 10:00:35 crc kubenswrapper[4744]: I1008 10:00:35.713995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" event={"ID":"e80d5289-7657-4835-8b43-936537129935","Type":"ContainerDied","Data":"aecd8b16e43f68379cfc120d16e7eb8185be6c508570bc5111b6f53bef6973a8"} Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.163728 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.358753 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-libvirt-secret-0\") pod \"e80d5289-7657-4835-8b43-936537129935\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.358843 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skrxn\" (UniqueName: \"kubernetes.io/projected/e80d5289-7657-4835-8b43-936537129935-kube-api-access-skrxn\") pod \"e80d5289-7657-4835-8b43-936537129935\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.358899 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-ssh-key\") pod \"e80d5289-7657-4835-8b43-936537129935\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.358938 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-libvirt-combined-ca-bundle\") pod \"e80d5289-7657-4835-8b43-936537129935\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.359003 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-ceph\") pod \"e80d5289-7657-4835-8b43-936537129935\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.359506 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-inventory\") pod \"e80d5289-7657-4835-8b43-936537129935\" (UID: \"e80d5289-7657-4835-8b43-936537129935\") " Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.364308 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e80d5289-7657-4835-8b43-936537129935-kube-api-access-skrxn" (OuterVolumeSpecName: "kube-api-access-skrxn") pod "e80d5289-7657-4835-8b43-936537129935" (UID: "e80d5289-7657-4835-8b43-936537129935"). InnerVolumeSpecName "kube-api-access-skrxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.364908 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-ceph" (OuterVolumeSpecName: "ceph") pod "e80d5289-7657-4835-8b43-936537129935" (UID: "e80d5289-7657-4835-8b43-936537129935"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.373313 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "e80d5289-7657-4835-8b43-936537129935" (UID: "e80d5289-7657-4835-8b43-936537129935"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.385337 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e80d5289-7657-4835-8b43-936537129935" (UID: "e80d5289-7657-4835-8b43-936537129935"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.385782 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "e80d5289-7657-4835-8b43-936537129935" (UID: "e80d5289-7657-4835-8b43-936537129935"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.394419 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-inventory" (OuterVolumeSpecName: "inventory") pod "e80d5289-7657-4835-8b43-936537129935" (UID: "e80d5289-7657-4835-8b43-936537129935"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.462095 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.462120 4744 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.462133 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skrxn\" (UniqueName: \"kubernetes.io/projected/e80d5289-7657-4835-8b43-936537129935-kube-api-access-skrxn\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.462142 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.462151 4744 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.462160 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e80d5289-7657-4835-8b43-936537129935-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.754828 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" event={"ID":"e80d5289-7657-4835-8b43-936537129935","Type":"ContainerDied","Data":"8a2ce474765a945e448fba4234681c7df5e4566495749482dbb252eee954cd60"} Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.754864 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a2ce474765a945e448fba4234681c7df5e4566495749482dbb252eee954cd60" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.754880 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-swsrc" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.863849 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z"] Oct 08 10:00:37 crc kubenswrapper[4744]: E1008 10:00:37.864278 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b2c49b4-9d9a-421a-8e26-644c75f84777" containerName="registry-server" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.864296 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b2c49b4-9d9a-421a-8e26-644c75f84777" containerName="registry-server" Oct 08 10:00:37 crc kubenswrapper[4744]: E1008 10:00:37.864314 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e80d5289-7657-4835-8b43-936537129935" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.864321 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e80d5289-7657-4835-8b43-936537129935" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 08 10:00:37 crc kubenswrapper[4744]: E1008 10:00:37.864331 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" containerName="registry-server" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.864337 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" containerName="registry-server" Oct 08 10:00:37 crc kubenswrapper[4744]: E1008 10:00:37.864351 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" containerName="extract-content" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.864356 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" containerName="extract-content" Oct 08 10:00:37 crc kubenswrapper[4744]: E1008 10:00:37.864392 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" containerName="extract-utilities" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.864399 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" containerName="extract-utilities" Oct 08 10:00:37 crc kubenswrapper[4744]: E1008 10:00:37.864408 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b2c49b4-9d9a-421a-8e26-644c75f84777" containerName="extract-utilities" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.864414 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b2c49b4-9d9a-421a-8e26-644c75f84777" containerName="extract-utilities" Oct 08 10:00:37 crc kubenswrapper[4744]: E1008 10:00:37.864421 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b2c49b4-9d9a-421a-8e26-644c75f84777" containerName="extract-content" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.864426 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b2c49b4-9d9a-421a-8e26-644c75f84777" containerName="extract-content" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.864612 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e80d5289-7657-4835-8b43-936537129935" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.864627 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b2c49b4-9d9a-421a-8e26-644c75f84777" containerName="registry-server" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.864645 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d08d26cc-7c0a-45d5-bbf2-70ad2405a6f6" containerName="registry-server" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.865249 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.869660 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.869702 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.869723 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.869665 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.869960 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mtxm4" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.870049 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.870189 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.870240 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.870264 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.923774 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z"] Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.973731 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.973774 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.973837 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.973867 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.973890 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.974094 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.974319 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.974352 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.974446 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvq42\" (UniqueName: \"kubernetes.io/projected/43d8aacb-0649-483b-9e07-bde175fd9d52-kube-api-access-nvq42\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.974499 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:37 crc kubenswrapper[4744]: I1008 10:00:37.974609 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/43d8aacb-0649-483b-9e07-bde175fd9d52-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.076453 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.076506 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.076542 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.076565 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.076583 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.076628 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.076677 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.076696 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.076717 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvq42\" (UniqueName: \"kubernetes.io/projected/43d8aacb-0649-483b-9e07-bde175fd9d52-kube-api-access-nvq42\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.076744 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.076790 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/43d8aacb-0649-483b-9e07-bde175fd9d52-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.077757 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/43d8aacb-0649-483b-9e07-bde175fd9d52-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.078474 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.081595 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.081801 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.082204 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.082891 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.084287 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.086142 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.091079 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.100287 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.100668 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvq42\" (UniqueName: \"kubernetes.io/projected/43d8aacb-0649-483b-9e07-bde175fd9d52-kube-api-access-nvq42\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.182841 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:00:38 crc kubenswrapper[4744]: I1008 10:00:38.798146 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z"] Oct 08 10:00:39 crc kubenswrapper[4744]: I1008 10:00:39.783782 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" event={"ID":"43d8aacb-0649-483b-9e07-bde175fd9d52","Type":"ContainerStarted","Data":"b0c7e3876e4f7a1932b6a8279feb9e229497db9d5b56adf3985239c9f261079e"} Oct 08 10:00:40 crc kubenswrapper[4744]: I1008 10:00:40.794742 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" event={"ID":"43d8aacb-0649-483b-9e07-bde175fd9d52","Type":"ContainerStarted","Data":"daecb685b47fd611f1a3c1189a27684e17aaf703239287db90dc0aee2fcd1102"} Oct 08 10:00:49 crc kubenswrapper[4744]: I1008 10:00:49.690243 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:00:49 crc kubenswrapper[4744]: I1008 10:00:49.691028 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:00:56 crc kubenswrapper[4744]: I1008 10:00:56.365977 4744 scope.go:117] "RemoveContainer" containerID="37a4f12941e6d9f3d3fef029d1a710111f71d8be62a0c60ed6e27260a915b35f" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.139936 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" podStartSLOduration=21.689280978 podStartE2EDuration="23.139913581s" podCreationTimestamp="2025-10-08 10:00:37 +0000 UTC" firstStartedPulling="2025-10-08 10:00:38.805125711 +0000 UTC m=+2934.052770950" lastFinishedPulling="2025-10-08 10:00:40.255758314 +0000 UTC m=+2935.503403553" observedRunningTime="2025-10-08 10:00:40.813242316 +0000 UTC m=+2936.060887555" watchObservedRunningTime="2025-10-08 10:01:00.139913581 +0000 UTC m=+2955.387558820" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.141931 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29331961-k29hs"] Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.143169 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.155514 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29331961-k29hs"] Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.279019 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-config-data\") pod \"keystone-cron-29331961-k29hs\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.279068 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s6nv\" (UniqueName: \"kubernetes.io/projected/6d3e0b57-177d-4351-8fce-3a758311bb35-kube-api-access-2s6nv\") pod \"keystone-cron-29331961-k29hs\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.279095 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-combined-ca-bundle\") pod \"keystone-cron-29331961-k29hs\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.279121 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-fernet-keys\") pod \"keystone-cron-29331961-k29hs\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.381213 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-config-data\") pod \"keystone-cron-29331961-k29hs\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.381258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s6nv\" (UniqueName: \"kubernetes.io/projected/6d3e0b57-177d-4351-8fce-3a758311bb35-kube-api-access-2s6nv\") pod \"keystone-cron-29331961-k29hs\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.381284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-combined-ca-bundle\") pod \"keystone-cron-29331961-k29hs\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.381310 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-fernet-keys\") pod \"keystone-cron-29331961-k29hs\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.387982 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-combined-ca-bundle\") pod \"keystone-cron-29331961-k29hs\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.387990 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-fernet-keys\") pod \"keystone-cron-29331961-k29hs\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.388677 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-config-data\") pod \"keystone-cron-29331961-k29hs\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.401756 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s6nv\" (UniqueName: \"kubernetes.io/projected/6d3e0b57-177d-4351-8fce-3a758311bb35-kube-api-access-2s6nv\") pod \"keystone-cron-29331961-k29hs\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.463947 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.931776 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29331961-k29hs"] Oct 08 10:01:00 crc kubenswrapper[4744]: I1008 10:01:00.976230 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331961-k29hs" event={"ID":"6d3e0b57-177d-4351-8fce-3a758311bb35","Type":"ContainerStarted","Data":"db433d7795938306c2a02761e8fc46cffe666ecb4f88d8f443d4046e323e1b11"} Oct 08 10:01:01 crc kubenswrapper[4744]: I1008 10:01:01.986153 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331961-k29hs" event={"ID":"6d3e0b57-177d-4351-8fce-3a758311bb35","Type":"ContainerStarted","Data":"505af279939534291acef6e3dd4d6074be90ac31e8d366bd2b52502f883d64f9"} Oct 08 10:01:02 crc kubenswrapper[4744]: I1008 10:01:02.004078 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29331961-k29hs" podStartSLOduration=2.00405873 podStartE2EDuration="2.00405873s" podCreationTimestamp="2025-10-08 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:01:01.999145684 +0000 UTC m=+2957.246790923" watchObservedRunningTime="2025-10-08 10:01:02.00405873 +0000 UTC m=+2957.251703969" Oct 08 10:01:05 crc kubenswrapper[4744]: I1008 10:01:05.012020 4744 generic.go:334] "Generic (PLEG): container finished" podID="6d3e0b57-177d-4351-8fce-3a758311bb35" containerID="505af279939534291acef6e3dd4d6074be90ac31e8d366bd2b52502f883d64f9" exitCode=0 Oct 08 10:01:05 crc kubenswrapper[4744]: I1008 10:01:05.012097 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331961-k29hs" event={"ID":"6d3e0b57-177d-4351-8fce-3a758311bb35","Type":"ContainerDied","Data":"505af279939534291acef6e3dd4d6074be90ac31e8d366bd2b52502f883d64f9"} Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.381360 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.491733 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-combined-ca-bundle\") pod \"6d3e0b57-177d-4351-8fce-3a758311bb35\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.491881 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s6nv\" (UniqueName: \"kubernetes.io/projected/6d3e0b57-177d-4351-8fce-3a758311bb35-kube-api-access-2s6nv\") pod \"6d3e0b57-177d-4351-8fce-3a758311bb35\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.491946 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-config-data\") pod \"6d3e0b57-177d-4351-8fce-3a758311bb35\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.492001 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-fernet-keys\") pod \"6d3e0b57-177d-4351-8fce-3a758311bb35\" (UID: \"6d3e0b57-177d-4351-8fce-3a758311bb35\") " Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.518609 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d3e0b57-177d-4351-8fce-3a758311bb35-kube-api-access-2s6nv" (OuterVolumeSpecName: "kube-api-access-2s6nv") pod "6d3e0b57-177d-4351-8fce-3a758311bb35" (UID: "6d3e0b57-177d-4351-8fce-3a758311bb35"). InnerVolumeSpecName "kube-api-access-2s6nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.519534 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6d3e0b57-177d-4351-8fce-3a758311bb35" (UID: "6d3e0b57-177d-4351-8fce-3a758311bb35"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.527595 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d3e0b57-177d-4351-8fce-3a758311bb35" (UID: "6d3e0b57-177d-4351-8fce-3a758311bb35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.548827 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-config-data" (OuterVolumeSpecName: "config-data") pod "6d3e0b57-177d-4351-8fce-3a758311bb35" (UID: "6d3e0b57-177d-4351-8fce-3a758311bb35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.595339 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.595543 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s6nv\" (UniqueName: \"kubernetes.io/projected/6d3e0b57-177d-4351-8fce-3a758311bb35-kube-api-access-2s6nv\") on node \"crc\" DevicePath \"\"" Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.595619 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:01:06 crc kubenswrapper[4744]: I1008 10:01:06.595704 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d3e0b57-177d-4351-8fce-3a758311bb35-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 10:01:07 crc kubenswrapper[4744]: I1008 10:01:07.035617 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29331961-k29hs" event={"ID":"6d3e0b57-177d-4351-8fce-3a758311bb35","Type":"ContainerDied","Data":"db433d7795938306c2a02761e8fc46cffe666ecb4f88d8f443d4046e323e1b11"} Oct 08 10:01:07 crc kubenswrapper[4744]: I1008 10:01:07.035671 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db433d7795938306c2a02761e8fc46cffe666ecb4f88d8f443d4046e323e1b11" Oct 08 10:01:07 crc kubenswrapper[4744]: I1008 10:01:07.035696 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29331961-k29hs" Oct 08 10:01:19 crc kubenswrapper[4744]: I1008 10:01:19.690059 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:01:19 crc kubenswrapper[4744]: I1008 10:01:19.690737 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:01:49 crc kubenswrapper[4744]: I1008 10:01:49.690338 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:01:49 crc kubenswrapper[4744]: I1008 10:01:49.690800 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:01:49 crc kubenswrapper[4744]: I1008 10:01:49.690841 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 10:01:49 crc kubenswrapper[4744]: I1008 10:01:49.691688 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 10:01:49 crc kubenswrapper[4744]: I1008 10:01:49.691740 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" gracePeriod=600 Oct 08 10:01:49 crc kubenswrapper[4744]: E1008 10:01:49.835316 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:01:50 crc kubenswrapper[4744]: I1008 10:01:50.401680 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" exitCode=0 Oct 08 10:01:50 crc kubenswrapper[4744]: I1008 10:01:50.401759 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8"} Oct 08 10:01:50 crc kubenswrapper[4744]: I1008 10:01:50.402075 4744 scope.go:117] "RemoveContainer" containerID="aa495b33f33b23e8dcca10511302304051a76ce3efd1ef1a213237865410ba57" Oct 08 10:01:50 crc kubenswrapper[4744]: I1008 10:01:50.402789 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:01:50 crc kubenswrapper[4744]: E1008 10:01:50.403081 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:02:03 crc kubenswrapper[4744]: I1008 10:02:03.454122 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:02:03 crc kubenswrapper[4744]: E1008 10:02:03.455280 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:02:15 crc kubenswrapper[4744]: I1008 10:02:15.459654 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:02:15 crc kubenswrapper[4744]: E1008 10:02:15.460341 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:02:26 crc kubenswrapper[4744]: I1008 10:02:26.453854 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:02:26 crc kubenswrapper[4744]: E1008 10:02:26.454699 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:02:40 crc kubenswrapper[4744]: I1008 10:02:40.453426 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:02:40 crc kubenswrapper[4744]: E1008 10:02:40.454291 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:02:51 crc kubenswrapper[4744]: I1008 10:02:51.453298 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:02:51 crc kubenswrapper[4744]: E1008 10:02:51.454293 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:03:02 crc kubenswrapper[4744]: I1008 10:03:02.453437 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:03:02 crc kubenswrapper[4744]: E1008 10:03:02.454178 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:03:14 crc kubenswrapper[4744]: I1008 10:03:14.454317 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:03:14 crc kubenswrapper[4744]: E1008 10:03:14.455799 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:03:29 crc kubenswrapper[4744]: I1008 10:03:29.453246 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:03:29 crc kubenswrapper[4744]: E1008 10:03:29.454053 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:03:41 crc kubenswrapper[4744]: I1008 10:03:41.453402 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:03:41 crc kubenswrapper[4744]: E1008 10:03:41.454325 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:03:53 crc kubenswrapper[4744]: I1008 10:03:53.453898 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:03:53 crc kubenswrapper[4744]: E1008 10:03:53.455184 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:04:08 crc kubenswrapper[4744]: I1008 10:04:08.454264 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:04:08 crc kubenswrapper[4744]: E1008 10:04:08.457156 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:04:20 crc kubenswrapper[4744]: I1008 10:04:20.453587 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:04:20 crc kubenswrapper[4744]: E1008 10:04:20.454298 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:04:33 crc kubenswrapper[4744]: I1008 10:04:33.453450 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:04:33 crc kubenswrapper[4744]: E1008 10:04:33.454210 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:04:36 crc kubenswrapper[4744]: I1008 10:04:36.822136 4744 generic.go:334] "Generic (PLEG): container finished" podID="43d8aacb-0649-483b-9e07-bde175fd9d52" containerID="daecb685b47fd611f1a3c1189a27684e17aaf703239287db90dc0aee2fcd1102" exitCode=0 Oct 08 10:04:36 crc kubenswrapper[4744]: I1008 10:04:36.822213 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" event={"ID":"43d8aacb-0649-483b-9e07-bde175fd9d52","Type":"ContainerDied","Data":"daecb685b47fd611f1a3c1189a27684e17aaf703239287db90dc0aee2fcd1102"} Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.293138 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.343915 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-migration-ssh-key-1\") pod \"43d8aacb-0649-483b-9e07-bde175fd9d52\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.343964 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-cell1-compute-config-1\") pod \"43d8aacb-0649-483b-9e07-bde175fd9d52\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.344044 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvq42\" (UniqueName: \"kubernetes.io/projected/43d8aacb-0649-483b-9e07-bde175fd9d52-kube-api-access-nvq42\") pod \"43d8aacb-0649-483b-9e07-bde175fd9d52\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.344085 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-ceph\") pod \"43d8aacb-0649-483b-9e07-bde175fd9d52\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.344117 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-cell1-compute-config-0\") pod \"43d8aacb-0649-483b-9e07-bde175fd9d52\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.344132 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-migration-ssh-key-0\") pod \"43d8aacb-0649-483b-9e07-bde175fd9d52\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.344240 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-inventory\") pod \"43d8aacb-0649-483b-9e07-bde175fd9d52\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.344259 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-ssh-key\") pod \"43d8aacb-0649-483b-9e07-bde175fd9d52\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.344279 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-custom-ceph-combined-ca-bundle\") pod \"43d8aacb-0649-483b-9e07-bde175fd9d52\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.344312 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/43d8aacb-0649-483b-9e07-bde175fd9d52-ceph-nova-0\") pod \"43d8aacb-0649-483b-9e07-bde175fd9d52\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.344335 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-extra-config-0\") pod \"43d8aacb-0649-483b-9e07-bde175fd9d52\" (UID: \"43d8aacb-0649-483b-9e07-bde175fd9d52\") " Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.358321 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43d8aacb-0649-483b-9e07-bde175fd9d52-kube-api-access-nvq42" (OuterVolumeSpecName: "kube-api-access-nvq42") pod "43d8aacb-0649-483b-9e07-bde175fd9d52" (UID: "43d8aacb-0649-483b-9e07-bde175fd9d52"). InnerVolumeSpecName "kube-api-access-nvq42". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.358526 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "43d8aacb-0649-483b-9e07-bde175fd9d52" (UID: "43d8aacb-0649-483b-9e07-bde175fd9d52"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.380190 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "43d8aacb-0649-483b-9e07-bde175fd9d52" (UID: "43d8aacb-0649-483b-9e07-bde175fd9d52"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.387928 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "43d8aacb-0649-483b-9e07-bde175fd9d52" (UID: "43d8aacb-0649-483b-9e07-bde175fd9d52"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.389775 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-inventory" (OuterVolumeSpecName: "inventory") pod "43d8aacb-0649-483b-9e07-bde175fd9d52" (UID: "43d8aacb-0649-483b-9e07-bde175fd9d52"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.394086 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "43d8aacb-0649-483b-9e07-bde175fd9d52" (UID: "43d8aacb-0649-483b-9e07-bde175fd9d52"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.395508 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-ceph" (OuterVolumeSpecName: "ceph") pod "43d8aacb-0649-483b-9e07-bde175fd9d52" (UID: "43d8aacb-0649-483b-9e07-bde175fd9d52"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.400615 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "43d8aacb-0649-483b-9e07-bde175fd9d52" (UID: "43d8aacb-0649-483b-9e07-bde175fd9d52"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.401022 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "43d8aacb-0649-483b-9e07-bde175fd9d52" (UID: "43d8aacb-0649-483b-9e07-bde175fd9d52"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.406912 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43d8aacb-0649-483b-9e07-bde175fd9d52-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "43d8aacb-0649-483b-9e07-bde175fd9d52" (UID: "43d8aacb-0649-483b-9e07-bde175fd9d52"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.414583 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "43d8aacb-0649-483b-9e07-bde175fd9d52" (UID: "43d8aacb-0649-483b-9e07-bde175fd9d52"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.447197 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvq42\" (UniqueName: \"kubernetes.io/projected/43d8aacb-0649-483b-9e07-bde175fd9d52-kube-api-access-nvq42\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.447239 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.447250 4744 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.447259 4744 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.447267 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.447276 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.447284 4744 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.447295 4744 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/43d8aacb-0649-483b-9e07-bde175fd9d52-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.447304 4744 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.447314 4744 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.447322 4744 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/43d8aacb-0649-483b-9e07-bde175fd9d52-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.855072 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" event={"ID":"43d8aacb-0649-483b-9e07-bde175fd9d52","Type":"ContainerDied","Data":"b0c7e3876e4f7a1932b6a8279feb9e229497db9d5b56adf3985239c9f261079e"} Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.855123 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0c7e3876e4f7a1932b6a8279feb9e229497db9d5b56adf3985239c9f261079e" Oct 08 10:04:38 crc kubenswrapper[4744]: I1008 10:04:38.855211 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z" Oct 08 10:04:48 crc kubenswrapper[4744]: I1008 10:04:48.452800 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:04:48 crc kubenswrapper[4744]: E1008 10:04:48.453641 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.572709 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 08 10:04:53 crc kubenswrapper[4744]: E1008 10:04:53.573783 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43d8aacb-0649-483b-9e07-bde175fd9d52" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.573805 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="43d8aacb-0649-483b-9e07-bde175fd9d52" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 08 10:04:53 crc kubenswrapper[4744]: E1008 10:04:53.573822 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d3e0b57-177d-4351-8fce-3a758311bb35" containerName="keystone-cron" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.573828 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d3e0b57-177d-4351-8fce-3a758311bb35" containerName="keystone-cron" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.574039 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d3e0b57-177d-4351-8fce-3a758311bb35" containerName="keystone-cron" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.574052 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="43d8aacb-0649-483b-9e07-bde175fd9d52" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.575303 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.578303 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.578608 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.595546 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.597440 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.600425 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.616975 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625240 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d4fd8637-8969-4c7f-9ecf-9276b5550484-ceph\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625298 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625333 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625504 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625524 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfkdl\" (UniqueName: \"kubernetes.io/projected/d4fd8637-8969-4c7f-9ecf-9276b5550484-kube-api-access-wfkdl\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625543 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-dev\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625566 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625590 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc1c714-af01-4205-8670-9055cdb3c623-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625611 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc1c714-af01-4205-8670-9055cdb3c623-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625641 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625666 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4fd8637-8969-4c7f-9ecf-9276b5550484-scripts\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625687 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625711 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625734 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc1c714-af01-4205-8670-9055cdb3c623-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625765 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4fd8637-8969-4c7f-9ecf-9276b5550484-config-data\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625789 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625815 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625833 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625854 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4fd8637-8969-4c7f-9ecf-9276b5550484-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625896 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4fd8637-8969-4c7f-9ecf-9276b5550484-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625917 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-run\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625941 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625968 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-run\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.625995 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-dev\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.626039 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.626074 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-lib-modules\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.626104 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9cc1c714-af01-4205-8670-9055cdb3c623-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.626129 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.626151 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-sys\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.626176 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-sys\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.626230 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4cmx\" (UniqueName: \"kubernetes.io/projected/9cc1c714-af01-4205-8670-9055cdb3c623-kube-api-access-f4cmx\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.626452 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9cc1c714-af01-4205-8670-9055cdb3c623-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.644865 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728348 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728487 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4fd8637-8969-4c7f-9ecf-9276b5550484-scripts\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728512 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728531 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728548 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc1c714-af01-4205-8670-9055cdb3c623-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728581 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4fd8637-8969-4c7f-9ecf-9276b5550484-config-data\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728601 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728620 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728638 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728653 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4fd8637-8969-4c7f-9ecf-9276b5550484-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728692 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4fd8637-8969-4c7f-9ecf-9276b5550484-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728709 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-run\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728722 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728741 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-run\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728763 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-dev\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728771 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728795 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728839 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728867 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-lib-modules\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9cc1c714-af01-4205-8670-9055cdb3c623-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728915 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728930 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728938 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728972 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-sys\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.728995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-sys\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729070 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4cmx\" (UniqueName: \"kubernetes.io/projected/9cc1c714-af01-4205-8670-9055cdb3c623-kube-api-access-f4cmx\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729088 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9cc1c714-af01-4205-8670-9055cdb3c623-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729140 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d4fd8637-8969-4c7f-9ecf-9276b5550484-ceph\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729162 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729201 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729199 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729220 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfkdl\" (UniqueName: \"kubernetes.io/projected/d4fd8637-8969-4c7f-9ecf-9276b5550484-kube-api-access-wfkdl\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729236 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-dev\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729199 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729253 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729271 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-lib-modules\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729278 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729346 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc1c714-af01-4205-8670-9055cdb3c623-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729417 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc1c714-af01-4205-8670-9055cdb3c623-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.730313 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.730360 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-sys\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.734486 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cc1c714-af01-4205-8670-9055cdb3c623-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.734696 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc1c714-af01-4205-8670-9055cdb3c623-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.734774 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-run\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.734803 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.736702 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.737114 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4fd8637-8969-4c7f-9ecf-9276b5550484-config-data\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.737142 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9cc1c714-af01-4205-8670-9055cdb3c623-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.737185 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-dev\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.740049 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d4fd8637-8969-4c7f-9ecf-9276b5550484-ceph\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.740046 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4fd8637-8969-4c7f-9ecf-9276b5550484-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.740117 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.740142 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-dev\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.740161 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-run\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.740309 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.729245 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-sys\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.743502 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9cc1c714-af01-4205-8670-9055cdb3c623-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.743641 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d4fd8637-8969-4c7f-9ecf-9276b5550484-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.744070 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc1c714-af01-4205-8670-9055cdb3c623-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.761065 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9cc1c714-af01-4205-8670-9055cdb3c623-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.761307 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4fd8637-8969-4c7f-9ecf-9276b5550484-scripts\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.761563 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4fd8637-8969-4c7f-9ecf-9276b5550484-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.765814 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfkdl\" (UniqueName: \"kubernetes.io/projected/d4fd8637-8969-4c7f-9ecf-9276b5550484-kube-api-access-wfkdl\") pod \"cinder-backup-0\" (UID: \"d4fd8637-8969-4c7f-9ecf-9276b5550484\") " pod="openstack/cinder-backup-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.765871 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4cmx\" (UniqueName: \"kubernetes.io/projected/9cc1c714-af01-4205-8670-9055cdb3c623-kube-api-access-f4cmx\") pod \"cinder-volume-volume1-0\" (UID: \"9cc1c714-af01-4205-8670-9055cdb3c623\") " pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.896279 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:53 crc kubenswrapper[4744]: I1008 10:04:53.918020 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.341139 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66fcbb9755-w2nbw"] Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.343046 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.355185 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.355361 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-xv72q" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.355599 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.355944 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.371773 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66fcbb9755-w2nbw"] Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.445325 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b492a22d-99c3-4ec6-8ab6-be3875679018-horizon-secret-key\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.445439 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b492a22d-99c3-4ec6-8ab6-be3875679018-logs\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.445510 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b492a22d-99c3-4ec6-8ab6-be3875679018-scripts\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.445554 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b492a22d-99c3-4ec6-8ab6-be3875679018-config-data\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.445599 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcd8v\" (UniqueName: \"kubernetes.io/projected/b492a22d-99c3-4ec6-8ab6-be3875679018-kube-api-access-wcd8v\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.456972 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.460238 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.474870 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.475063 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.475174 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lwftp" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.475286 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.539554 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:04:54 crc kubenswrapper[4744]: E1008 10:04:54.540236 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run kube-api-access-fwdkl logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[ceph combined-ca-bundle config-data glance httpd-run kube-api-access-fwdkl logs public-tls-certs scripts]: context canceled" pod="openstack/glance-default-external-api-0" podUID="996c71b5-a09d-4e94-9ddf-2691531ba808" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.547223 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b492a22d-99c3-4ec6-8ab6-be3875679018-scripts\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.547292 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b492a22d-99c3-4ec6-8ab6-be3875679018-config-data\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.547350 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcd8v\" (UniqueName: \"kubernetes.io/projected/b492a22d-99c3-4ec6-8ab6-be3875679018-kube-api-access-wcd8v\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.547404 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b492a22d-99c3-4ec6-8ab6-be3875679018-horizon-secret-key\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.547462 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b492a22d-99c3-4ec6-8ab6-be3875679018-logs\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.547903 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b492a22d-99c3-4ec6-8ab6-be3875679018-logs\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.548994 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b492a22d-99c3-4ec6-8ab6-be3875679018-scripts\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.549908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b492a22d-99c3-4ec6-8ab6-be3875679018-config-data\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.559446 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.561073 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.574996 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.578203 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b492a22d-99c3-4ec6-8ab6-be3875679018-horizon-secret-key\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.595588 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.595696 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.609689 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcd8v\" (UniqueName: \"kubernetes.io/projected/b492a22d-99c3-4ec6-8ab6-be3875679018-kube-api-access-wcd8v\") pod \"horizon-66fcbb9755-w2nbw\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.613057 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.649391 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/996c71b5-a09d-4e94-9ddf-2691531ba808-logs\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.649426 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/996c71b5-a09d-4e94-9ddf-2691531ba808-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.649445 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-scripts\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.649479 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-config-data\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.649510 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.649539 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.649587 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.649621 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/996c71b5-a09d-4e94-9ddf-2691531ba808-ceph\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.649646 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwdkl\" (UniqueName: \"kubernetes.io/projected/996c71b5-a09d-4e94-9ddf-2691531ba808-kube-api-access-fwdkl\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.669791 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.672125 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-v7c88"] Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.673215 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v7c88" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.684006 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-v7c88"] Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753047 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/996c71b5-a09d-4e94-9ddf-2691531ba808-ceph\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753080 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwdkl\" (UniqueName: \"kubernetes.io/projected/996c71b5-a09d-4e94-9ddf-2691531ba808-kube-api-access-fwdkl\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753122 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753173 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753215 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7f486a8-44cc-4847-ba8d-8096b24ce808-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753248 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7f486a8-44cc-4847-ba8d-8096b24ce808-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753268 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e7f486a8-44cc-4847-ba8d-8096b24ce808-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753287 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753312 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/996c71b5-a09d-4e94-9ddf-2691531ba808-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753329 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/996c71b5-a09d-4e94-9ddf-2691531ba808-logs\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753346 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-scripts\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753378 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9nks\" (UniqueName: \"kubernetes.io/projected/e7f486a8-44cc-4847-ba8d-8096b24ce808-kube-api-access-l9nks\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753409 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-config-data\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753431 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753463 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753481 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753508 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlc7m\" (UniqueName: \"kubernetes.io/projected/119d6cdd-c521-4ad3-b4c0-ceb7eb956281-kube-api-access-vlc7m\") pod \"manila-db-create-v7c88\" (UID: \"119d6cdd-c521-4ad3-b4c0-ceb7eb956281\") " pod="openstack/manila-db-create-v7c88" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753539 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.753556 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.757804 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.762061 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/996c71b5-a09d-4e94-9ddf-2691531ba808-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.762932 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/996c71b5-a09d-4e94-9ddf-2691531ba808-logs\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.776133 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-config-data\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.787352 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/996c71b5-a09d-4e94-9ddf-2691531ba808-ceph\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.787515 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.791190 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.797432 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:04:54 crc kubenswrapper[4744]: E1008 10:04:54.810179 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-l9nks logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="e7f486a8-44cc-4847-ba8d-8096b24ce808" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.832136 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwdkl\" (UniqueName: \"kubernetes.io/projected/996c71b5-a09d-4e94-9ddf-2691531ba808-kube-api-access-fwdkl\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.834140 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-scripts\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.896067 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.896128 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlc7m\" (UniqueName: \"kubernetes.io/projected/119d6cdd-c521-4ad3-b4c0-ceb7eb956281-kube-api-access-vlc7m\") pod \"manila-db-create-v7c88\" (UID: \"119d6cdd-c521-4ad3-b4c0-ceb7eb956281\") " pod="openstack/manila-db-create-v7c88" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.896228 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.896324 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.896480 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.896559 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7f486a8-44cc-4847-ba8d-8096b24ce808-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.896585 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7f486a8-44cc-4847-ba8d-8096b24ce808-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.896609 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e7f486a8-44cc-4847-ba8d-8096b24ce808-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.896629 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.896687 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9nks\" (UniqueName: \"kubernetes.io/projected/e7f486a8-44cc-4847-ba8d-8096b24ce808-kube-api-access-l9nks\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.897227 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.897478 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.897851 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7f486a8-44cc-4847-ba8d-8096b24ce808-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.898117 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7f486a8-44cc-4847-ba8d-8096b24ce808-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.915312 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5f866cb5ff-dkrbz"] Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.916826 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.923417 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.924751 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e7f486a8-44cc-4847-ba8d-8096b24ce808-ceph\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.925568 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.946469 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlc7m\" (UniqueName: \"kubernetes.io/projected/119d6cdd-c521-4ad3-b4c0-ceb7eb956281-kube-api-access-vlc7m\") pod \"manila-db-create-v7c88\" (UID: \"119d6cdd-c521-4ad3-b4c0-ceb7eb956281\") " pod="openstack/manila-db-create-v7c88" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.952411 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f866cb5ff-dkrbz"] Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.971600 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.975631 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.978631 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.981248 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9nks\" (UniqueName: \"kubernetes.io/projected/e7f486a8-44cc-4847-ba8d-8096b24ce808-kube-api-access-l9nks\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.997488 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v7c88" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.999096 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-scripts\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.999619 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-horizon-secret-key\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.999849 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-config-data\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:54 crc kubenswrapper[4744]: I1008 10:04:54.999937 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-logs\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.017835 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl6sz\" (UniqueName: \"kubernetes.io/projected/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-kube-api-access-vl6sz\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.123334 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-horizon-secret-key\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.126412 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-config-data\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.126535 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-logs\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.126785 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl6sz\" (UniqueName: \"kubernetes.io/projected/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-kube-api-access-vl6sz\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.126969 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-scripts\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.127885 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-config-data\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.128163 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-logs\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.128939 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-scripts\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.172414 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.172643 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"d4fd8637-8969-4c7f-9ecf-9276b5550484","Type":"ContainerStarted","Data":"d386d1e44f83fb79540d8a5f9d5c2884a1bbacb5925e8c3a8553ff603937e65a"} Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.172901 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.183925 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.204227 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.204496 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl6sz\" (UniqueName: \"kubernetes.io/projected/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-kube-api-access-vl6sz\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.207574 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.221840 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-horizon-secret-key\") pod \"horizon-5f866cb5ff-dkrbz\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.324963 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.330324 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"996c71b5-a09d-4e94-9ddf-2691531ba808\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.330471 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-config-data\") pod \"996c71b5-a09d-4e94-9ddf-2691531ba808\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.330524 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-config-data\") pod \"e7f486a8-44cc-4847-ba8d-8096b24ce808\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.331074 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-combined-ca-bundle\") pod \"996c71b5-a09d-4e94-9ddf-2691531ba808\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.331120 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-scripts\") pod \"e7f486a8-44cc-4847-ba8d-8096b24ce808\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.331153 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-combined-ca-bundle\") pod \"e7f486a8-44cc-4847-ba8d-8096b24ce808\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.331203 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7f486a8-44cc-4847-ba8d-8096b24ce808-httpd-run\") pod \"e7f486a8-44cc-4847-ba8d-8096b24ce808\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.331679 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f486a8-44cc-4847-ba8d-8096b24ce808-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e7f486a8-44cc-4847-ba8d-8096b24ce808" (UID: "e7f486a8-44cc-4847-ba8d-8096b24ce808"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.331944 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-internal-tls-certs\") pod \"e7f486a8-44cc-4847-ba8d-8096b24ce808\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.331991 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-scripts\") pod \"996c71b5-a09d-4e94-9ddf-2691531ba808\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.332009 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9nks\" (UniqueName: \"kubernetes.io/projected/e7f486a8-44cc-4847-ba8d-8096b24ce808-kube-api-access-l9nks\") pod \"e7f486a8-44cc-4847-ba8d-8096b24ce808\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.332038 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7f486a8-44cc-4847-ba8d-8096b24ce808-logs\") pod \"e7f486a8-44cc-4847-ba8d-8096b24ce808\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.332075 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwdkl\" (UniqueName: \"kubernetes.io/projected/996c71b5-a09d-4e94-9ddf-2691531ba808-kube-api-access-fwdkl\") pod \"996c71b5-a09d-4e94-9ddf-2691531ba808\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.332094 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-public-tls-certs\") pod \"996c71b5-a09d-4e94-9ddf-2691531ba808\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.332111 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/996c71b5-a09d-4e94-9ddf-2691531ba808-ceph\") pod \"996c71b5-a09d-4e94-9ddf-2691531ba808\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.332170 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/996c71b5-a09d-4e94-9ddf-2691531ba808-httpd-run\") pod \"996c71b5-a09d-4e94-9ddf-2691531ba808\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.332191 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"e7f486a8-44cc-4847-ba8d-8096b24ce808\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.332255 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e7f486a8-44cc-4847-ba8d-8096b24ce808-ceph\") pod \"e7f486a8-44cc-4847-ba8d-8096b24ce808\" (UID: \"e7f486a8-44cc-4847-ba8d-8096b24ce808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.332320 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/996c71b5-a09d-4e94-9ddf-2691531ba808-logs\") pod \"996c71b5-a09d-4e94-9ddf-2691531ba808\" (UID: \"996c71b5-a09d-4e94-9ddf-2691531ba808\") " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.332704 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7f486a8-44cc-4847-ba8d-8096b24ce808-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.334345 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/996c71b5-a09d-4e94-9ddf-2691531ba808-logs" (OuterVolumeSpecName: "logs") pod "996c71b5-a09d-4e94-9ddf-2691531ba808" (UID: "996c71b5-a09d-4e94-9ddf-2691531ba808"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.335819 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f486a8-44cc-4847-ba8d-8096b24ce808-logs" (OuterVolumeSpecName: "logs") pod "e7f486a8-44cc-4847-ba8d-8096b24ce808" (UID: "e7f486a8-44cc-4847-ba8d-8096b24ce808"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.342871 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "996c71b5-a09d-4e94-9ddf-2691531ba808" (UID: "996c71b5-a09d-4e94-9ddf-2691531ba808"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.343881 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f486a8-44cc-4847-ba8d-8096b24ce808-kube-api-access-l9nks" (OuterVolumeSpecName: "kube-api-access-l9nks") pod "e7f486a8-44cc-4847-ba8d-8096b24ce808" (UID: "e7f486a8-44cc-4847-ba8d-8096b24ce808"). InnerVolumeSpecName "kube-api-access-l9nks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.343899 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-config-data" (OuterVolumeSpecName: "config-data") pod "996c71b5-a09d-4e94-9ddf-2691531ba808" (UID: "996c71b5-a09d-4e94-9ddf-2691531ba808"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.344775 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-scripts" (OuterVolumeSpecName: "scripts") pod "e7f486a8-44cc-4847-ba8d-8096b24ce808" (UID: "e7f486a8-44cc-4847-ba8d-8096b24ce808"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.346130 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/996c71b5-a09d-4e94-9ddf-2691531ba808-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "996c71b5-a09d-4e94-9ddf-2691531ba808" (UID: "996c71b5-a09d-4e94-9ddf-2691531ba808"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.359784 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "996c71b5-a09d-4e94-9ddf-2691531ba808" (UID: "996c71b5-a09d-4e94-9ddf-2691531ba808"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.367602 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f486a8-44cc-4847-ba8d-8096b24ce808-ceph" (OuterVolumeSpecName: "ceph") pod "e7f486a8-44cc-4847-ba8d-8096b24ce808" (UID: "e7f486a8-44cc-4847-ba8d-8096b24ce808"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.370523 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "e7f486a8-44cc-4847-ba8d-8096b24ce808" (UID: "e7f486a8-44cc-4847-ba8d-8096b24ce808"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.370730 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/996c71b5-a09d-4e94-9ddf-2691531ba808-kube-api-access-fwdkl" (OuterVolumeSpecName: "kube-api-access-fwdkl") pod "996c71b5-a09d-4e94-9ddf-2691531ba808" (UID: "996c71b5-a09d-4e94-9ddf-2691531ba808"). InnerVolumeSpecName "kube-api-access-fwdkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.372397 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-scripts" (OuterVolumeSpecName: "scripts") pod "996c71b5-a09d-4e94-9ddf-2691531ba808" (UID: "996c71b5-a09d-4e94-9ddf-2691531ba808"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.372512 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-config-data" (OuterVolumeSpecName: "config-data") pod "e7f486a8-44cc-4847-ba8d-8096b24ce808" (UID: "e7f486a8-44cc-4847-ba8d-8096b24ce808"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.372539 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7f486a8-44cc-4847-ba8d-8096b24ce808" (UID: "e7f486a8-44cc-4847-ba8d-8096b24ce808"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.372599 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/996c71b5-a09d-4e94-9ddf-2691531ba808-ceph" (OuterVolumeSpecName: "ceph") pod "996c71b5-a09d-4e94-9ddf-2691531ba808" (UID: "996c71b5-a09d-4e94-9ddf-2691531ba808"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.372689 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e7f486a8-44cc-4847-ba8d-8096b24ce808" (UID: "e7f486a8-44cc-4847-ba8d-8096b24ce808"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.379059 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "996c71b5-a09d-4e94-9ddf-2691531ba808" (UID: "996c71b5-a09d-4e94-9ddf-2691531ba808"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.435956 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.435990 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9nks\" (UniqueName: \"kubernetes.io/projected/e7f486a8-44cc-4847-ba8d-8096b24ce808-kube-api-access-l9nks\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436000 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7f486a8-44cc-4847-ba8d-8096b24ce808-logs\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436010 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwdkl\" (UniqueName: \"kubernetes.io/projected/996c71b5-a09d-4e94-9ddf-2691531ba808-kube-api-access-fwdkl\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436020 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436030 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/996c71b5-a09d-4e94-9ddf-2691531ba808-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436037 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/996c71b5-a09d-4e94-9ddf-2691531ba808-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436063 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436071 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e7f486a8-44cc-4847-ba8d-8096b24ce808-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436079 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/996c71b5-a09d-4e94-9ddf-2691531ba808-logs\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436091 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436100 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436108 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436117 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/996c71b5-a09d-4e94-9ddf-2691531ba808-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436126 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436133 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.436141 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7f486a8-44cc-4847-ba8d-8096b24ce808-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.443983 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.469719 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.473734 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.543637 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.543666 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.612199 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66fcbb9755-w2nbw"] Oct 08 10:04:55 crc kubenswrapper[4744]: I1008 10:04:55.820112 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-v7c88"] Oct 08 10:04:55 crc kubenswrapper[4744]: W1008 10:04:55.971827 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod119d6cdd_c521_4ad3_b4c0_ceb7eb956281.slice/crio-ffb06ab62a5574c742f5d931c4f226e28b2183fc3a7d17edde9aabcbab7f5b58 WatchSource:0}: Error finding container ffb06ab62a5574c742f5d931c4f226e28b2183fc3a7d17edde9aabcbab7f5b58: Status 404 returned error can't find the container with id ffb06ab62a5574c742f5d931c4f226e28b2183fc3a7d17edde9aabcbab7f5b58 Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.003804 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f866cb5ff-dkrbz"] Oct 08 10:04:56 crc kubenswrapper[4744]: W1008 10:04:56.061780 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b41d1ed_6119_42a5_8ee7_d4e820d595a8.slice/crio-6a7604bd17f7c309ce0d185ad6662276f2cd1383d3ecfc3d6977189019080e68 WatchSource:0}: Error finding container 6a7604bd17f7c309ce0d185ad6662276f2cd1383d3ecfc3d6977189019080e68: Status 404 returned error can't find the container with id 6a7604bd17f7c309ce0d185ad6662276f2cd1383d3ecfc3d6977189019080e68 Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.194516 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v7c88" event={"ID":"119d6cdd-c521-4ad3-b4c0-ceb7eb956281","Type":"ContainerStarted","Data":"ffb06ab62a5574c742f5d931c4f226e28b2183fc3a7d17edde9aabcbab7f5b58"} Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.205348 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f866cb5ff-dkrbz" event={"ID":"8b41d1ed-6119-42a5-8ee7-d4e820d595a8","Type":"ContainerStarted","Data":"6a7604bd17f7c309ce0d185ad6662276f2cd1383d3ecfc3d6977189019080e68"} Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.208511 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"9cc1c714-af01-4205-8670-9055cdb3c623","Type":"ContainerStarted","Data":"92ebeef8e3a60e0dd6d3c08ba2a48932656c7fab57273db37aee00e43b3c62b9"} Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.214308 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.215174 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66fcbb9755-w2nbw" event={"ID":"b492a22d-99c3-4ec6-8ab6-be3875679018","Type":"ContainerStarted","Data":"fd5ab7654834adf9e2c03431cb47aee23483bb16c06f10ce19ea1c78fdadddf7"} Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.215218 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.305976 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.318075 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.336387 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.339475 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.343752 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.343965 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.344183 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lwftp" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.345303 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.377147 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.418594 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.433290 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.469453 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-scripts\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.469531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2wg7\" (UniqueName: \"kubernetes.io/projected/934f5902-4baa-4ad4-912c-dc1e83724cb7-kube-api-access-f2wg7\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.469572 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-config-data\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.469713 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.469733 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/934f5902-4baa-4ad4-912c-dc1e83724cb7-logs\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.469812 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/934f5902-4baa-4ad4-912c-dc1e83724cb7-ceph\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.469846 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/934f5902-4baa-4ad4-912c-dc1e83724cb7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.469942 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.469982 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.502249 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.505875 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.512736 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.514538 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.527791 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.572007 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-config-data\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.572213 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.572256 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/934f5902-4baa-4ad4-912c-dc1e83724cb7-logs\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.572951 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/934f5902-4baa-4ad4-912c-dc1e83724cb7-logs\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.582146 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.584325 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-config-data\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.584904 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/934f5902-4baa-4ad4-912c-dc1e83724cb7-ceph\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.572365 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/934f5902-4baa-4ad4-912c-dc1e83724cb7-ceph\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.588613 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/934f5902-4baa-4ad4-912c-dc1e83724cb7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.588964 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.589092 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.589179 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-scripts\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.589218 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2wg7\" (UniqueName: \"kubernetes.io/projected/934f5902-4baa-4ad4-912c-dc1e83724cb7-kube-api-access-f2wg7\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.589772 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.594016 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/934f5902-4baa-4ad4-912c-dc1e83724cb7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.597440 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-scripts\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.607297 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.617694 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2wg7\" (UniqueName: \"kubernetes.io/projected/934f5902-4baa-4ad4-912c-dc1e83724cb7-kube-api-access-f2wg7\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.677207 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.697302 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2xwl\" (UniqueName: \"kubernetes.io/projected/7e842955-068f-4e7e-830d-f64c8ce81dbd-kube-api-access-m2xwl\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.697475 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.697570 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.697636 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e842955-068f-4e7e-830d-f64c8ce81dbd-logs\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.697733 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.697947 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7e842955-068f-4e7e-830d-f64c8ce81dbd-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.697984 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.698064 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.698135 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e842955-068f-4e7e-830d-f64c8ce81dbd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.708802 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.803905 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2xwl\" (UniqueName: \"kubernetes.io/projected/7e842955-068f-4e7e-830d-f64c8ce81dbd-kube-api-access-m2xwl\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.804615 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.804705 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.804785 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e842955-068f-4e7e-830d-f64c8ce81dbd-logs\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.804875 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.805022 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7e842955-068f-4e7e-830d-f64c8ce81dbd-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.805112 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.805213 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.805299 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e842955-068f-4e7e-830d-f64c8ce81dbd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.805851 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e842955-068f-4e7e-830d-f64c8ce81dbd-logs\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.806167 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e842955-068f-4e7e-830d-f64c8ce81dbd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.806469 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.813954 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7e842955-068f-4e7e-830d-f64c8ce81dbd-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.814203 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.814270 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.814793 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.832211 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.845254 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2xwl\" (UniqueName: \"kubernetes.io/projected/7e842955-068f-4e7e-830d-f64c8ce81dbd-kube-api-access-m2xwl\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:56 crc kubenswrapper[4744]: I1008 10:04:56.871963 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:04:57 crc kubenswrapper[4744]: I1008 10:04:57.179762 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 10:04:57 crc kubenswrapper[4744]: I1008 10:04:57.243991 4744 generic.go:334] "Generic (PLEG): container finished" podID="119d6cdd-c521-4ad3-b4c0-ceb7eb956281" containerID="46c76ad342f1f057138dd6fac2a75bd62c797e077d5ed791bed275fbce6b4bd5" exitCode=0 Oct 08 10:04:57 crc kubenswrapper[4744]: I1008 10:04:57.244103 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v7c88" event={"ID":"119d6cdd-c521-4ad3-b4c0-ceb7eb956281","Type":"ContainerDied","Data":"46c76ad342f1f057138dd6fac2a75bd62c797e077d5ed791bed275fbce6b4bd5"} Oct 08 10:04:57 crc kubenswrapper[4744]: I1008 10:04:57.259142 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"d4fd8637-8969-4c7f-9ecf-9276b5550484","Type":"ContainerStarted","Data":"807ea46e0acb3cc215c1f4e20013ff25a20ac97d8f8b7379fb5215f911b11e32"} Oct 08 10:04:57 crc kubenswrapper[4744]: I1008 10:04:57.259177 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"d4fd8637-8969-4c7f-9ecf-9276b5550484","Type":"ContainerStarted","Data":"c49f47ada45d6de80fe8bc420944642e4ce13e9e7d8e249627bd8dbcc253e10a"} Oct 08 10:04:57 crc kubenswrapper[4744]: I1008 10:04:57.308304 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.275846132 podStartE2EDuration="4.308286073s" podCreationTimestamp="2025-10-08 10:04:53 +0000 UTC" firstStartedPulling="2025-10-08 10:04:55.033880706 +0000 UTC m=+3190.281525945" lastFinishedPulling="2025-10-08 10:04:56.066320647 +0000 UTC m=+3191.313965886" observedRunningTime="2025-10-08 10:04:57.296697995 +0000 UTC m=+3192.544343224" watchObservedRunningTime="2025-10-08 10:04:57.308286073 +0000 UTC m=+3192.555931312" Oct 08 10:04:57 crc kubenswrapper[4744]: I1008 10:04:57.312698 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"9cc1c714-af01-4205-8670-9055cdb3c623","Type":"ContainerStarted","Data":"ef71b1e955ace400c86640cc51f463c13787f1afd89da73002c53a8f16ed19ce"} Oct 08 10:04:57 crc kubenswrapper[4744]: I1008 10:04:57.491776 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="996c71b5-a09d-4e94-9ddf-2691531ba808" path="/var/lib/kubelet/pods/996c71b5-a09d-4e94-9ddf-2691531ba808/volumes" Oct 08 10:04:57 crc kubenswrapper[4744]: I1008 10:04:57.493636 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7f486a8-44cc-4847-ba8d-8096b24ce808" path="/var/lib/kubelet/pods/e7f486a8-44cc-4847-ba8d-8096b24ce808/volumes" Oct 08 10:04:57 crc kubenswrapper[4744]: I1008 10:04:57.612178 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:04:57 crc kubenswrapper[4744]: W1008 10:04:57.666471 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod934f5902_4baa_4ad4_912c_dc1e83724cb7.slice/crio-740b45c6212829c653a10be6602395892cd21352bce87cc4d819602cfff38a0e WatchSource:0}: Error finding container 740b45c6212829c653a10be6602395892cd21352bce87cc4d819602cfff38a0e: Status 404 returned error can't find the container with id 740b45c6212829c653a10be6602395892cd21352bce87cc4d819602cfff38a0e Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.088498 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f866cb5ff-dkrbz"] Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.329356 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f8f98d57b-66wrm"] Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.345930 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.398725 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.405785 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f8f98d57b-66wrm"] Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.407296 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.425181 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"9cc1c714-af01-4205-8670-9055cdb3c623","Type":"ContainerStarted","Data":"5c8a4a54d4bef7f08ccefdab49a651ab89426483e39f036db2953aa827a922d2"} Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.431513 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"934f5902-4baa-4ad4-912c-dc1e83724cb7","Type":"ContainerStarted","Data":"740b45c6212829c653a10be6602395892cd21352bce87cc4d819602cfff38a0e"} Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.474200 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49150e8b-4e4b-467f-ae7b-a6d14b63554d-logs\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.474406 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-horizon-tls-certs\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.474438 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbgfv\" (UniqueName: \"kubernetes.io/projected/49150e8b-4e4b-467f-ae7b-a6d14b63554d-kube-api-access-hbgfv\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.474485 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-horizon-secret-key\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.474568 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49150e8b-4e4b-467f-ae7b-a6d14b63554d-scripts\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.474593 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-combined-ca-bundle\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.474665 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49150e8b-4e4b-467f-ae7b-a6d14b63554d-config-data\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.567629 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=4.325369992 podStartE2EDuration="5.567613176s" podCreationTimestamp="2025-10-08 10:04:53 +0000 UTC" firstStartedPulling="2025-10-08 10:04:55.389615541 +0000 UTC m=+3190.637260780" lastFinishedPulling="2025-10-08 10:04:56.631858725 +0000 UTC m=+3191.879503964" observedRunningTime="2025-10-08 10:04:58.558413573 +0000 UTC m=+3193.806058802" watchObservedRunningTime="2025-10-08 10:04:58.567613176 +0000 UTC m=+3193.815258415" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.579926 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49150e8b-4e4b-467f-ae7b-a6d14b63554d-config-data\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.580787 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49150e8b-4e4b-467f-ae7b-a6d14b63554d-logs\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.580950 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-horizon-tls-certs\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.580997 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbgfv\" (UniqueName: \"kubernetes.io/projected/49150e8b-4e4b-467f-ae7b-a6d14b63554d-kube-api-access-hbgfv\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.581016 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-horizon-secret-key\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.581090 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49150e8b-4e4b-467f-ae7b-a6d14b63554d-scripts\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.581116 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-combined-ca-bundle\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.581154 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49150e8b-4e4b-467f-ae7b-a6d14b63554d-config-data\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.582959 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49150e8b-4e4b-467f-ae7b-a6d14b63554d-logs\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.586306 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49150e8b-4e4b-467f-ae7b-a6d14b63554d-scripts\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.600978 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-horizon-secret-key\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.613161 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-horizon-tls-certs\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.641241 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-combined-ca-bundle\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.661145 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66fcbb9755-w2nbw"] Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.682544 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbgfv\" (UniqueName: \"kubernetes.io/projected/49150e8b-4e4b-467f-ae7b-a6d14b63554d-kube-api-access-hbgfv\") pod \"horizon-7f8f98d57b-66wrm\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.711979 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.739887 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.825520 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5549798486-rbbgm"] Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.886271 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.903332 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.931010 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 08 10:04:58 crc kubenswrapper[4744]: I1008 10:04:58.982062 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.045935 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5549798486-rbbgm"] Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.088610 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f337fa5f-1088-407d-a832-856ff504cd72-logs\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.088950 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f337fa5f-1088-407d-a832-856ff504cd72-horizon-secret-key\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.088991 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f337fa5f-1088-407d-a832-856ff504cd72-config-data\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.089025 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtqzb\" (UniqueName: \"kubernetes.io/projected/f337fa5f-1088-407d-a832-856ff504cd72-kube-api-access-xtqzb\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.089069 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f337fa5f-1088-407d-a832-856ff504cd72-horizon-tls-certs\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.089122 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f337fa5f-1088-407d-a832-856ff504cd72-scripts\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.089148 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f337fa5f-1088-407d-a832-856ff504cd72-combined-ca-bundle\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.196543 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f337fa5f-1088-407d-a832-856ff504cd72-scripts\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.196597 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f337fa5f-1088-407d-a832-856ff504cd72-combined-ca-bundle\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.196670 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f337fa5f-1088-407d-a832-856ff504cd72-logs\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.196698 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f337fa5f-1088-407d-a832-856ff504cd72-horizon-secret-key\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.196731 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f337fa5f-1088-407d-a832-856ff504cd72-config-data\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.196760 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtqzb\" (UniqueName: \"kubernetes.io/projected/f337fa5f-1088-407d-a832-856ff504cd72-kube-api-access-xtqzb\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.196810 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f337fa5f-1088-407d-a832-856ff504cd72-horizon-tls-certs\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.227743 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f337fa5f-1088-407d-a832-856ff504cd72-horizon-secret-key\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.227842 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f337fa5f-1088-407d-a832-856ff504cd72-combined-ca-bundle\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.229135 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f337fa5f-1088-407d-a832-856ff504cd72-horizon-tls-certs\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.245908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f337fa5f-1088-407d-a832-856ff504cd72-logs\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.246622 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f337fa5f-1088-407d-a832-856ff504cd72-scripts\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.248476 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f337fa5f-1088-407d-a832-856ff504cd72-config-data\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.302403 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtqzb\" (UniqueName: \"kubernetes.io/projected/f337fa5f-1088-407d-a832-856ff504cd72-kube-api-access-xtqzb\") pod \"horizon-5549798486-rbbgm\" (UID: \"f337fa5f-1088-407d-a832-856ff504cd72\") " pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.384078 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.427893 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v7c88" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.471057 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-v7c88" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.532534 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-v7c88" event={"ID":"119d6cdd-c521-4ad3-b4c0-ceb7eb956281","Type":"ContainerDied","Data":"ffb06ab62a5574c742f5d931c4f226e28b2183fc3a7d17edde9aabcbab7f5b58"} Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.532952 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffb06ab62a5574c742f5d931c4f226e28b2183fc3a7d17edde9aabcbab7f5b58" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.532966 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7e842955-068f-4e7e-830d-f64c8ce81dbd","Type":"ContainerStarted","Data":"9dd7ab893f8ee66c229d93ee0ed5991303726d8f39b23f82db96302e95c4cae8"} Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.621294 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlc7m\" (UniqueName: \"kubernetes.io/projected/119d6cdd-c521-4ad3-b4c0-ceb7eb956281-kube-api-access-vlc7m\") pod \"119d6cdd-c521-4ad3-b4c0-ceb7eb956281\" (UID: \"119d6cdd-c521-4ad3-b4c0-ceb7eb956281\") " Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.632863 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/119d6cdd-c521-4ad3-b4c0-ceb7eb956281-kube-api-access-vlc7m" (OuterVolumeSpecName: "kube-api-access-vlc7m") pod "119d6cdd-c521-4ad3-b4c0-ceb7eb956281" (UID: "119d6cdd-c521-4ad3-b4c0-ceb7eb956281"). InnerVolumeSpecName "kube-api-access-vlc7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:04:59 crc kubenswrapper[4744]: I1008 10:04:59.727574 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlc7m\" (UniqueName: \"kubernetes.io/projected/119d6cdd-c521-4ad3-b4c0-ceb7eb956281-kube-api-access-vlc7m\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:00 crc kubenswrapper[4744]: I1008 10:05:00.146737 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f8f98d57b-66wrm"] Oct 08 10:05:00 crc kubenswrapper[4744]: I1008 10:05:00.452921 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:05:00 crc kubenswrapper[4744]: E1008 10:05:00.453473 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:05:00 crc kubenswrapper[4744]: I1008 10:05:00.532522 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f8f98d57b-66wrm" event={"ID":"49150e8b-4e4b-467f-ae7b-a6d14b63554d","Type":"ContainerStarted","Data":"0212bbd8f0391c84b8a5e18e4f583c8b159752be2f8fc1075d40ddb0783a5029"} Oct 08 10:05:00 crc kubenswrapper[4744]: I1008 10:05:00.535766 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"934f5902-4baa-4ad4-912c-dc1e83724cb7","Type":"ContainerStarted","Data":"46d183dd4af5e22613d76bef02e426b2b1413d5f84fa49a289a5dbd8d31d1ebb"} Oct 08 10:05:00 crc kubenswrapper[4744]: I1008 10:05:00.559400 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5549798486-rbbgm"] Oct 08 10:05:01 crc kubenswrapper[4744]: I1008 10:05:01.554671 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"934f5902-4baa-4ad4-912c-dc1e83724cb7","Type":"ContainerStarted","Data":"fdeff26fcb4500efd4ffed76f7af12dc963c9ad10b168bf02287719f5302a532"} Oct 08 10:05:01 crc kubenswrapper[4744]: I1008 10:05:01.555912 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="934f5902-4baa-4ad4-912c-dc1e83724cb7" containerName="glance-log" containerID="cri-o://46d183dd4af5e22613d76bef02e426b2b1413d5f84fa49a289a5dbd8d31d1ebb" gracePeriod=30 Oct 08 10:05:01 crc kubenswrapper[4744]: I1008 10:05:01.556579 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="934f5902-4baa-4ad4-912c-dc1e83724cb7" containerName="glance-httpd" containerID="cri-o://fdeff26fcb4500efd4ffed76f7af12dc963c9ad10b168bf02287719f5302a532" gracePeriod=30 Oct 08 10:05:01 crc kubenswrapper[4744]: I1008 10:05:01.566399 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5549798486-rbbgm" event={"ID":"f337fa5f-1088-407d-a832-856ff504cd72","Type":"ContainerStarted","Data":"f0ebbcaeaf53a0770c9544510d74668fdc55b253e9c058ec9b31a2182d93c179"} Oct 08 10:05:01 crc kubenswrapper[4744]: I1008 10:05:01.577315 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7e842955-068f-4e7e-830d-f64c8ce81dbd","Type":"ContainerStarted","Data":"53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282"} Oct 08 10:05:01 crc kubenswrapper[4744]: I1008 10:05:01.577452 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7e842955-068f-4e7e-830d-f64c8ce81dbd","Type":"ContainerStarted","Data":"3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c"} Oct 08 10:05:01 crc kubenswrapper[4744]: I1008 10:05:01.577618 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7e842955-068f-4e7e-830d-f64c8ce81dbd" containerName="glance-log" containerID="cri-o://3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c" gracePeriod=30 Oct 08 10:05:01 crc kubenswrapper[4744]: I1008 10:05:01.577718 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7e842955-068f-4e7e-830d-f64c8ce81dbd" containerName="glance-httpd" containerID="cri-o://53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282" gracePeriod=30 Oct 08 10:05:01 crc kubenswrapper[4744]: I1008 10:05:01.630674 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.630655097 podStartE2EDuration="5.630655097s" podCreationTimestamp="2025-10-08 10:04:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:05:01.598527187 +0000 UTC m=+3196.846172426" watchObservedRunningTime="2025-10-08 10:05:01.630655097 +0000 UTC m=+3196.878300336" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.519804 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.528188 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-config-data\") pod \"7e842955-068f-4e7e-830d-f64c8ce81dbd\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.528235 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2xwl\" (UniqueName: \"kubernetes.io/projected/7e842955-068f-4e7e-830d-f64c8ce81dbd-kube-api-access-m2xwl\") pod \"7e842955-068f-4e7e-830d-f64c8ce81dbd\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.528261 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-internal-tls-certs\") pod \"7e842955-068f-4e7e-830d-f64c8ce81dbd\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.528296 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7e842955-068f-4e7e-830d-f64c8ce81dbd-ceph\") pod \"7e842955-068f-4e7e-830d-f64c8ce81dbd\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.528333 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e842955-068f-4e7e-830d-f64c8ce81dbd-logs\") pod \"7e842955-068f-4e7e-830d-f64c8ce81dbd\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.528401 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-scripts\") pod \"7e842955-068f-4e7e-830d-f64c8ce81dbd\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.528450 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-combined-ca-bundle\") pod \"7e842955-068f-4e7e-830d-f64c8ce81dbd\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.528495 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e842955-068f-4e7e-830d-f64c8ce81dbd-httpd-run\") pod \"7e842955-068f-4e7e-830d-f64c8ce81dbd\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.528611 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"7e842955-068f-4e7e-830d-f64c8ce81dbd\" (UID: \"7e842955-068f-4e7e-830d-f64c8ce81dbd\") " Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.531774 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e842955-068f-4e7e-830d-f64c8ce81dbd-logs" (OuterVolumeSpecName: "logs") pod "7e842955-068f-4e7e-830d-f64c8ce81dbd" (UID: "7e842955-068f-4e7e-830d-f64c8ce81dbd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.532893 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e842955-068f-4e7e-830d-f64c8ce81dbd-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7e842955-068f-4e7e-830d-f64c8ce81dbd" (UID: "7e842955-068f-4e7e-830d-f64c8ce81dbd"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.542435 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "7e842955-068f-4e7e-830d-f64c8ce81dbd" (UID: "7e842955-068f-4e7e-830d-f64c8ce81dbd"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.546017 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-scripts" (OuterVolumeSpecName: "scripts") pod "7e842955-068f-4e7e-830d-f64c8ce81dbd" (UID: "7e842955-068f-4e7e-830d-f64c8ce81dbd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.554923 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e842955-068f-4e7e-830d-f64c8ce81dbd-kube-api-access-m2xwl" (OuterVolumeSpecName: "kube-api-access-m2xwl") pod "7e842955-068f-4e7e-830d-f64c8ce81dbd" (UID: "7e842955-068f-4e7e-830d-f64c8ce81dbd"). InnerVolumeSpecName "kube-api-access-m2xwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.572618 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e842955-068f-4e7e-830d-f64c8ce81dbd-ceph" (OuterVolumeSpecName: "ceph") pod "7e842955-068f-4e7e-830d-f64c8ce81dbd" (UID: "7e842955-068f-4e7e-830d-f64c8ce81dbd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.611468 4744 generic.go:334] "Generic (PLEG): container finished" podID="7e842955-068f-4e7e-830d-f64c8ce81dbd" containerID="53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282" exitCode=143 Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.611516 4744 generic.go:334] "Generic (PLEG): container finished" podID="7e842955-068f-4e7e-830d-f64c8ce81dbd" containerID="3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c" exitCode=143 Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.611575 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7e842955-068f-4e7e-830d-f64c8ce81dbd","Type":"ContainerDied","Data":"53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282"} Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.611624 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7e842955-068f-4e7e-830d-f64c8ce81dbd","Type":"ContainerDied","Data":"3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c"} Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.611639 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7e842955-068f-4e7e-830d-f64c8ce81dbd","Type":"ContainerDied","Data":"9dd7ab893f8ee66c229d93ee0ed5991303726d8f39b23f82db96302e95c4cae8"} Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.611663 4744 scope.go:117] "RemoveContainer" containerID="53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.611993 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.619891 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e842955-068f-4e7e-830d-f64c8ce81dbd" (UID: "7e842955-068f-4e7e-830d-f64c8ce81dbd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.621792 4744 generic.go:334] "Generic (PLEG): container finished" podID="934f5902-4baa-4ad4-912c-dc1e83724cb7" containerID="fdeff26fcb4500efd4ffed76f7af12dc963c9ad10b168bf02287719f5302a532" exitCode=143 Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.621949 4744 generic.go:334] "Generic (PLEG): container finished" podID="934f5902-4baa-4ad4-912c-dc1e83724cb7" containerID="46d183dd4af5e22613d76bef02e426b2b1413d5f84fa49a289a5dbd8d31d1ebb" exitCode=143 Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.622065 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"934f5902-4baa-4ad4-912c-dc1e83724cb7","Type":"ContainerDied","Data":"fdeff26fcb4500efd4ffed76f7af12dc963c9ad10b168bf02287719f5302a532"} Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.622483 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"934f5902-4baa-4ad4-912c-dc1e83724cb7","Type":"ContainerDied","Data":"46d183dd4af5e22613d76bef02e426b2b1413d5f84fa49a289a5dbd8d31d1ebb"} Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.633360 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.633525 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2xwl\" (UniqueName: \"kubernetes.io/projected/7e842955-068f-4e7e-830d-f64c8ce81dbd-kube-api-access-m2xwl\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.633549 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7e842955-068f-4e7e-830d-f64c8ce81dbd-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.633558 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7e842955-068f-4e7e-830d-f64c8ce81dbd-logs\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.633566 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.633575 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.633583 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7e842955-068f-4e7e-830d-f64c8ce81dbd-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.652926 4744 scope.go:117] "RemoveContainer" containerID="3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.657152 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-config-data" (OuterVolumeSpecName: "config-data") pod "7e842955-068f-4e7e-830d-f64c8ce81dbd" (UID: "7e842955-068f-4e7e-830d-f64c8ce81dbd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.666018 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.667111 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7e842955-068f-4e7e-830d-f64c8ce81dbd" (UID: "7e842955-068f-4e7e-830d-f64c8ce81dbd"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.693774 4744 scope.go:117] "RemoveContainer" containerID="53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282" Oct 08 10:05:02 crc kubenswrapper[4744]: E1008 10:05:02.694213 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282\": container with ID starting with 53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282 not found: ID does not exist" containerID="53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.694347 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282"} err="failed to get container status \"53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282\": rpc error: code = NotFound desc = could not find container \"53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282\": container with ID starting with 53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282 not found: ID does not exist" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.694402 4744 scope.go:117] "RemoveContainer" containerID="3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c" Oct 08 10:05:02 crc kubenswrapper[4744]: E1008 10:05:02.694823 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c\": container with ID starting with 3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c not found: ID does not exist" containerID="3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.694857 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c"} err="failed to get container status \"3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c\": rpc error: code = NotFound desc = could not find container \"3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c\": container with ID starting with 3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c not found: ID does not exist" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.694882 4744 scope.go:117] "RemoveContainer" containerID="53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.695177 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282"} err="failed to get container status \"53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282\": rpc error: code = NotFound desc = could not find container \"53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282\": container with ID starting with 53a5b33a2d2265531938af6934fd91496a72660d98e5ff27baaadb8687773282 not found: ID does not exist" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.695197 4744 scope.go:117] "RemoveContainer" containerID="3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.695621 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c"} err="failed to get container status \"3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c\": rpc error: code = NotFound desc = could not find container \"3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c\": container with ID starting with 3b26dcc2ba7d0d0689f3703c8d26910a6960e7449482f46ba74c09316cd0854c not found: ID does not exist" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.736352 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.736408 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.736419 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e842955-068f-4e7e-830d-f64c8ce81dbd-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.980510 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:05:02 crc kubenswrapper[4744]: I1008 10:05:02.998710 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.030786 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:05:03 crc kubenswrapper[4744]: E1008 10:05:03.031205 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="119d6cdd-c521-4ad3-b4c0-ceb7eb956281" containerName="mariadb-database-create" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.031227 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="119d6cdd-c521-4ad3-b4c0-ceb7eb956281" containerName="mariadb-database-create" Oct 08 10:05:03 crc kubenswrapper[4744]: E1008 10:05:03.031258 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e842955-068f-4e7e-830d-f64c8ce81dbd" containerName="glance-httpd" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.031266 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e842955-068f-4e7e-830d-f64c8ce81dbd" containerName="glance-httpd" Oct 08 10:05:03 crc kubenswrapper[4744]: E1008 10:05:03.031286 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e842955-068f-4e7e-830d-f64c8ce81dbd" containerName="glance-log" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.031293 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e842955-068f-4e7e-830d-f64c8ce81dbd" containerName="glance-log" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.031524 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="119d6cdd-c521-4ad3-b4c0-ceb7eb956281" containerName="mariadb-database-create" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.031553 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e842955-068f-4e7e-830d-f64c8ce81dbd" containerName="glance-httpd" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.031565 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e842955-068f-4e7e-830d-f64c8ce81dbd" containerName="glance-log" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.032750 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.041987 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.042365 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.052018 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.060023 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bdd463c-e43b-4740-ba42-3981cb6b1380-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.060109 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bdd463c-e43b-4740-ba42-3981cb6b1380-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.060137 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.060160 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bdd463c-e43b-4740-ba42-3981cb6b1380-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.060183 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bdd463c-e43b-4740-ba42-3981cb6b1380-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.060307 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2bdd463c-e43b-4740-ba42-3981cb6b1380-ceph\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.060393 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlmjr\" (UniqueName: \"kubernetes.io/projected/2bdd463c-e43b-4740-ba42-3981cb6b1380-kube-api-access-qlmjr\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.060508 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bdd463c-e43b-4740-ba42-3981cb6b1380-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.060542 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdd463c-e43b-4740-ba42-3981cb6b1380-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.161698 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2bdd463c-e43b-4740-ba42-3981cb6b1380-ceph\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.161752 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlmjr\" (UniqueName: \"kubernetes.io/projected/2bdd463c-e43b-4740-ba42-3981cb6b1380-kube-api-access-qlmjr\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.161812 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bdd463c-e43b-4740-ba42-3981cb6b1380-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.161844 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdd463c-e43b-4740-ba42-3981cb6b1380-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.161912 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bdd463c-e43b-4740-ba42-3981cb6b1380-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.161984 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bdd463c-e43b-4740-ba42-3981cb6b1380-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.162011 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.162029 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bdd463c-e43b-4740-ba42-3981cb6b1380-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.162050 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bdd463c-e43b-4740-ba42-3981cb6b1380-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.164084 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bdd463c-e43b-4740-ba42-3981cb6b1380-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.166599 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.167705 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bdd463c-e43b-4740-ba42-3981cb6b1380-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.170873 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdd463c-e43b-4740-ba42-3981cb6b1380-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.170912 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bdd463c-e43b-4740-ba42-3981cb6b1380-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.181676 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bdd463c-e43b-4740-ba42-3981cb6b1380-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.188391 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2bdd463c-e43b-4740-ba42-3981cb6b1380-ceph\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.192784 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bdd463c-e43b-4740-ba42-3981cb6b1380-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.202673 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.213459 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlmjr\" (UniqueName: \"kubernetes.io/projected/2bdd463c-e43b-4740-ba42-3981cb6b1380-kube-api-access-qlmjr\") pod \"glance-default-internal-api-0\" (UID: \"2bdd463c-e43b-4740-ba42-3981cb6b1380\") " pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.370517 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.382604 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.470456 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-config-data\") pod \"934f5902-4baa-4ad4-912c-dc1e83724cb7\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.470547 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/934f5902-4baa-4ad4-912c-dc1e83724cb7-logs\") pod \"934f5902-4baa-4ad4-912c-dc1e83724cb7\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.470636 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/934f5902-4baa-4ad4-912c-dc1e83724cb7-httpd-run\") pod \"934f5902-4baa-4ad4-912c-dc1e83724cb7\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.470685 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-combined-ca-bundle\") pod \"934f5902-4baa-4ad4-912c-dc1e83724cb7\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.470729 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-public-tls-certs\") pod \"934f5902-4baa-4ad4-912c-dc1e83724cb7\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.470772 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"934f5902-4baa-4ad4-912c-dc1e83724cb7\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.470816 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-scripts\") pod \"934f5902-4baa-4ad4-912c-dc1e83724cb7\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.470851 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2wg7\" (UniqueName: \"kubernetes.io/projected/934f5902-4baa-4ad4-912c-dc1e83724cb7-kube-api-access-f2wg7\") pod \"934f5902-4baa-4ad4-912c-dc1e83724cb7\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.470880 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/934f5902-4baa-4ad4-912c-dc1e83724cb7-ceph\") pod \"934f5902-4baa-4ad4-912c-dc1e83724cb7\" (UID: \"934f5902-4baa-4ad4-912c-dc1e83724cb7\") " Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.478214 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/934f5902-4baa-4ad4-912c-dc1e83724cb7-logs" (OuterVolumeSpecName: "logs") pod "934f5902-4baa-4ad4-912c-dc1e83724cb7" (UID: "934f5902-4baa-4ad4-912c-dc1e83724cb7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.478550 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/934f5902-4baa-4ad4-912c-dc1e83724cb7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "934f5902-4baa-4ad4-912c-dc1e83724cb7" (UID: "934f5902-4baa-4ad4-912c-dc1e83724cb7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.486861 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/934f5902-4baa-4ad4-912c-dc1e83724cb7-ceph" (OuterVolumeSpecName: "ceph") pod "934f5902-4baa-4ad4-912c-dc1e83724cb7" (UID: "934f5902-4baa-4ad4-912c-dc1e83724cb7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.490695 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "934f5902-4baa-4ad4-912c-dc1e83724cb7" (UID: "934f5902-4baa-4ad4-912c-dc1e83724cb7"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.502696 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-scripts" (OuterVolumeSpecName: "scripts") pod "934f5902-4baa-4ad4-912c-dc1e83724cb7" (UID: "934f5902-4baa-4ad4-912c-dc1e83724cb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.505550 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/934f5902-4baa-4ad4-912c-dc1e83724cb7-kube-api-access-f2wg7" (OuterVolumeSpecName: "kube-api-access-f2wg7") pod "934f5902-4baa-4ad4-912c-dc1e83724cb7" (UID: "934f5902-4baa-4ad4-912c-dc1e83724cb7"). InnerVolumeSpecName "kube-api-access-f2wg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.534553 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "934f5902-4baa-4ad4-912c-dc1e83724cb7" (UID: "934f5902-4baa-4ad4-912c-dc1e83724cb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.537558 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e842955-068f-4e7e-830d-f64c8ce81dbd" path="/var/lib/kubelet/pods/7e842955-068f-4e7e-830d-f64c8ce81dbd/volumes" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.565150 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "934f5902-4baa-4ad4-912c-dc1e83724cb7" (UID: "934f5902-4baa-4ad4-912c-dc1e83724cb7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.573605 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.573645 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2wg7\" (UniqueName: \"kubernetes.io/projected/934f5902-4baa-4ad4-912c-dc1e83724cb7-kube-api-access-f2wg7\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.573661 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/934f5902-4baa-4ad4-912c-dc1e83724cb7-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.573671 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/934f5902-4baa-4ad4-912c-dc1e83724cb7-logs\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.573680 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/934f5902-4baa-4ad4-912c-dc1e83724cb7-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.573691 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.573705 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.573730 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.632560 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-config-data" (OuterVolumeSpecName: "config-data") pod "934f5902-4baa-4ad4-912c-dc1e83724cb7" (UID: "934f5902-4baa-4ad4-912c-dc1e83724cb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.653185 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.681594 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.681618 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/934f5902-4baa-4ad4-912c-dc1e83724cb7-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.696093 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"934f5902-4baa-4ad4-912c-dc1e83724cb7","Type":"ContainerDied","Data":"740b45c6212829c653a10be6602395892cd21352bce87cc4d819602cfff38a0e"} Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.696171 4744 scope.go:117] "RemoveContainer" containerID="fdeff26fcb4500efd4ffed76f7af12dc963c9ad10b168bf02287719f5302a532" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.696331 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.829508 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.884424 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.917456 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:05:03 crc kubenswrapper[4744]: E1008 10:05:03.917847 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="934f5902-4baa-4ad4-912c-dc1e83724cb7" containerName="glance-log" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.917858 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="934f5902-4baa-4ad4-912c-dc1e83724cb7" containerName="glance-log" Oct 08 10:05:03 crc kubenswrapper[4744]: E1008 10:05:03.917884 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="934f5902-4baa-4ad4-912c-dc1e83724cb7" containerName="glance-httpd" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.917890 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="934f5902-4baa-4ad4-912c-dc1e83724cb7" containerName="glance-httpd" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.918052 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="934f5902-4baa-4ad4-912c-dc1e83724cb7" containerName="glance-log" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.918073 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="934f5902-4baa-4ad4-912c-dc1e83724cb7" containerName="glance-httpd" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.919053 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.935482 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.936207 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.969636 4744 scope.go:117] "RemoveContainer" containerID="46d183dd4af5e22613d76bef02e426b2b1413d5f84fa49a289a5dbd8d31d1ebb" Oct 08 10:05:03 crc kubenswrapper[4744]: I1008 10:05:03.975008 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.107514 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26c9e730-4ef5-41f3-b861-fbe778351306-scripts\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.112024 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26c9e730-4ef5-41f3-b861-fbe778351306-config-data\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.112181 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hthc4\" (UniqueName: \"kubernetes.io/projected/26c9e730-4ef5-41f3-b861-fbe778351306-kube-api-access-hthc4\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.112383 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26c9e730-4ef5-41f3-b861-fbe778351306-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.112493 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26c9e730-4ef5-41f3-b861-fbe778351306-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.112634 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26c9e730-4ef5-41f3-b861-fbe778351306-logs\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.112750 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26c9e730-4ef5-41f3-b861-fbe778351306-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.112796 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.112838 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/26c9e730-4ef5-41f3-b861-fbe778351306-ceph\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.215872 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26c9e730-4ef5-41f3-b861-fbe778351306-logs\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.215971 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26c9e730-4ef5-41f3-b861-fbe778351306-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.216010 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.216038 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/26c9e730-4ef5-41f3-b861-fbe778351306-ceph\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.216104 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26c9e730-4ef5-41f3-b861-fbe778351306-scripts\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.216149 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26c9e730-4ef5-41f3-b861-fbe778351306-config-data\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.216197 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hthc4\" (UniqueName: \"kubernetes.io/projected/26c9e730-4ef5-41f3-b861-fbe778351306-kube-api-access-hthc4\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.216276 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26c9e730-4ef5-41f3-b861-fbe778351306-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.216310 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26c9e730-4ef5-41f3-b861-fbe778351306-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.216635 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26c9e730-4ef5-41f3-b861-fbe778351306-logs\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.217914 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26c9e730-4ef5-41f3-b861-fbe778351306-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.218161 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.229531 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/26c9e730-4ef5-41f3-b861-fbe778351306-ceph\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.233283 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26c9e730-4ef5-41f3-b861-fbe778351306-scripts\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.233359 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26c9e730-4ef5-41f3-b861-fbe778351306-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.242832 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/26c9e730-4ef5-41f3-b861-fbe778351306-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.244026 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26c9e730-4ef5-41f3-b861-fbe778351306-config-data\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.272711 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hthc4\" (UniqueName: \"kubernetes.io/projected/26c9e730-4ef5-41f3-b861-fbe778351306-kube-api-access-hthc4\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.392553 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"26c9e730-4ef5-41f3-b861-fbe778351306\") " pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.479227 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.562448 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.652870 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.739678 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bdd463c-e43b-4740-ba42-3981cb6b1380","Type":"ContainerStarted","Data":"30bd12bd51e9c1c3ea64e8f205067cfa38b9a8a2cc076abe3ce1f5cc86ae18d1"} Oct 08 10:05:04 crc kubenswrapper[4744]: I1008 10:05:04.896729 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-volume-volume1-0" podUID="9cc1c714-af01-4205-8670-9055cdb3c623" containerName="cinder-volume" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 10:05:05 crc kubenswrapper[4744]: I1008 10:05:05.404232 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 10:05:05 crc kubenswrapper[4744]: I1008 10:05:05.494741 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="934f5902-4baa-4ad4-912c-dc1e83724cb7" path="/var/lib/kubelet/pods/934f5902-4baa-4ad4-912c-dc1e83724cb7/volumes" Oct 08 10:05:05 crc kubenswrapper[4744]: I1008 10:05:05.773203 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"26c9e730-4ef5-41f3-b861-fbe778351306","Type":"ContainerStarted","Data":"69202e6c0ee29d5cc6444d05bee00ba32dc070dee6907ae1b68e98bf439e347f"} Oct 08 10:05:06 crc kubenswrapper[4744]: I1008 10:05:06.850465 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bdd463c-e43b-4740-ba42-3981cb6b1380","Type":"ContainerStarted","Data":"7e7aad8d68af3cc3a7900df48b13960da0d18cbd46299bc551d1711e05f470ca"} Oct 08 10:05:06 crc kubenswrapper[4744]: I1008 10:05:06.855907 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"26c9e730-4ef5-41f3-b861-fbe778351306","Type":"ContainerStarted","Data":"f68547590debc887dda08404e3c4e310423bc06681437d0ea44969fd905159d4"} Oct 08 10:05:07 crc kubenswrapper[4744]: I1008 10:05:07.894092 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bdd463c-e43b-4740-ba42-3981cb6b1380","Type":"ContainerStarted","Data":"780ccc0a119c5de6408057fffe0ba5550ded9e54627d588891d282374f87c82d"} Oct 08 10:05:07 crc kubenswrapper[4744]: I1008 10:05:07.933152 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.933113578 podStartE2EDuration="5.933113578s" podCreationTimestamp="2025-10-08 10:05:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:05:07.923927037 +0000 UTC m=+3203.171572276" watchObservedRunningTime="2025-10-08 10:05:07.933113578 +0000 UTC m=+3203.180758817" Oct 08 10:05:08 crc kubenswrapper[4744]: I1008 10:05:08.911817 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 08 10:05:08 crc kubenswrapper[4744]: I1008 10:05:08.913442 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"26c9e730-4ef5-41f3-b861-fbe778351306","Type":"ContainerStarted","Data":"c8c91472bbc13dce87bf8ff783736165c0830a3a26ec66c47f7d92eb9d72a909"} Oct 08 10:05:09 crc kubenswrapper[4744]: I1008 10:05:09.006831 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.006788719 podStartE2EDuration="6.006788719s" podCreationTimestamp="2025-10-08 10:05:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:05:08.985700412 +0000 UTC m=+3204.233345661" watchObservedRunningTime="2025-10-08 10:05:09.006788719 +0000 UTC m=+3204.254433978" Oct 08 10:05:13 crc kubenswrapper[4744]: I1008 10:05:13.371427 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:13 crc kubenswrapper[4744]: I1008 10:05:13.372191 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:13 crc kubenswrapper[4744]: I1008 10:05:13.413155 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:13 crc kubenswrapper[4744]: I1008 10:05:13.439107 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:13 crc kubenswrapper[4744]: I1008 10:05:13.966613 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:13 crc kubenswrapper[4744]: I1008 10:05:13.966656 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.454003 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:05:14 crc kubenswrapper[4744]: E1008 10:05:14.454238 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.566633 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.567307 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.631025 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.670850 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.751638 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-5c7c-account-create-ww5bj"] Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.761927 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-5c7c-account-create-ww5bj" Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.765836 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.785729 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-5c7c-account-create-ww5bj"] Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.825682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmt4s\" (UniqueName: \"kubernetes.io/projected/308c5a41-6d01-4684-9ab0-a23e0e372252-kube-api-access-tmt4s\") pod \"manila-5c7c-account-create-ww5bj\" (UID: \"308c5a41-6d01-4684-9ab0-a23e0e372252\") " pod="openstack/manila-5c7c-account-create-ww5bj" Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.928945 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmt4s\" (UniqueName: \"kubernetes.io/projected/308c5a41-6d01-4684-9ab0-a23e0e372252-kube-api-access-tmt4s\") pod \"manila-5c7c-account-create-ww5bj\" (UID: \"308c5a41-6d01-4684-9ab0-a23e0e372252\") " pod="openstack/manila-5c7c-account-create-ww5bj" Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.956012 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmt4s\" (UniqueName: \"kubernetes.io/projected/308c5a41-6d01-4684-9ab0-a23e0e372252-kube-api-access-tmt4s\") pod \"manila-5c7c-account-create-ww5bj\" (UID: \"308c5a41-6d01-4684-9ab0-a23e0e372252\") " pod="openstack/manila-5c7c-account-create-ww5bj" Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.979576 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f866cb5ff-dkrbz" event={"ID":"8b41d1ed-6119-42a5-8ee7-d4e820d595a8","Type":"ContainerStarted","Data":"411b8a7ef38838dd31b1a9be5935619da8e930b2e99f4869d6776712e32d926c"} Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.981438 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66fcbb9755-w2nbw" event={"ID":"b492a22d-99c3-4ec6-8ab6-be3875679018","Type":"ContainerStarted","Data":"9d230f0533ead7fc718808c57ac2da9650713e9e95833d6f29715c32d1a60b99"} Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.982226 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 10:05:14 crc kubenswrapper[4744]: I1008 10:05:14.982520 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 10:05:15 crc kubenswrapper[4744]: I1008 10:05:15.108423 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-5c7c-account-create-ww5bj" Oct 08 10:05:15 crc kubenswrapper[4744]: I1008 10:05:15.731310 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-5c7c-account-create-ww5bj"] Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.063506 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-5c7c-account-create-ww5bj" event={"ID":"308c5a41-6d01-4684-9ab0-a23e0e372252","Type":"ContainerStarted","Data":"15480379b42e33d485eb87ce235ba093fbab8f9c51083fc60881170d4d9f3114"} Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.073597 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5549798486-rbbgm" event={"ID":"f337fa5f-1088-407d-a832-856ff504cd72","Type":"ContainerStarted","Data":"787cc59928f8321815c9e32b08a66e999d06f5e7345824086b491d99d0b2101e"} Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.073640 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5549798486-rbbgm" event={"ID":"f337fa5f-1088-407d-a832-856ff504cd72","Type":"ContainerStarted","Data":"75afa79935526646550e6965e10f438acd33576c0dd99e8eddedbe71b65e4b1a"} Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.079909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f8f98d57b-66wrm" event={"ID":"49150e8b-4e4b-467f-ae7b-a6d14b63554d","Type":"ContainerStarted","Data":"2139c1999ee135e2b62d6e294e9ec99ea5173c7f974b9cfaa2786d28f6dac0bb"} Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.079950 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f8f98d57b-66wrm" event={"ID":"49150e8b-4e4b-467f-ae7b-a6d14b63554d","Type":"ContainerStarted","Data":"621e50ea58c51b02bda6c75857ac031b6d770e3776c4efe36a0937f13140b393"} Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.085039 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66fcbb9755-w2nbw" event={"ID":"b492a22d-99c3-4ec6-8ab6-be3875679018","Type":"ContainerStarted","Data":"6961c20b15cbd567c5c14acb0c8c3f875a72fb0094f0bdf4fa1888723670f461"} Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.085215 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66fcbb9755-w2nbw" podUID="b492a22d-99c3-4ec6-8ab6-be3875679018" containerName="horizon-log" containerID="cri-o://9d230f0533ead7fc718808c57ac2da9650713e9e95833d6f29715c32d1a60b99" gracePeriod=30 Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.085522 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66fcbb9755-w2nbw" podUID="b492a22d-99c3-4ec6-8ab6-be3875679018" containerName="horizon" containerID="cri-o://6961c20b15cbd567c5c14acb0c8c3f875a72fb0094f0bdf4fa1888723670f461" gracePeriod=30 Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.093490 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5f866cb5ff-dkrbz" podUID="8b41d1ed-6119-42a5-8ee7-d4e820d595a8" containerName="horizon-log" containerID="cri-o://411b8a7ef38838dd31b1a9be5935619da8e930b2e99f4869d6776712e32d926c" gracePeriod=30 Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.093603 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5f866cb5ff-dkrbz" podUID="8b41d1ed-6119-42a5-8ee7-d4e820d595a8" containerName="horizon" containerID="cri-o://bf53ea7e94ab1a9dc1f1252d25d0883df0ff4a2fae485a02e28790d7e8156bff" gracePeriod=30 Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.093664 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f866cb5ff-dkrbz" event={"ID":"8b41d1ed-6119-42a5-8ee7-d4e820d595a8","Type":"ContainerStarted","Data":"bf53ea7e94ab1a9dc1f1252d25d0883df0ff4a2fae485a02e28790d7e8156bff"} Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.093723 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.093731 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.123177 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5549798486-rbbgm" podStartSLOduration=4.179322312 podStartE2EDuration="18.123159278s" podCreationTimestamp="2025-10-08 10:04:58 +0000 UTC" firstStartedPulling="2025-10-08 10:05:00.615814479 +0000 UTC m=+3195.863459718" lastFinishedPulling="2025-10-08 10:05:14.559651445 +0000 UTC m=+3209.807296684" observedRunningTime="2025-10-08 10:05:16.099177181 +0000 UTC m=+3211.346822430" watchObservedRunningTime="2025-10-08 10:05:16.123159278 +0000 UTC m=+3211.370804517" Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.155886 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5f866cb5ff-dkrbz" podStartSLOduration=3.673759845 podStartE2EDuration="22.155871705s" podCreationTimestamp="2025-10-08 10:04:54 +0000 UTC" firstStartedPulling="2025-10-08 10:04:56.063907611 +0000 UTC m=+3191.311552860" lastFinishedPulling="2025-10-08 10:05:14.546019481 +0000 UTC m=+3209.793664720" observedRunningTime="2025-10-08 10:05:16.144122833 +0000 UTC m=+3211.391768092" watchObservedRunningTime="2025-10-08 10:05:16.155871705 +0000 UTC m=+3211.403516934" Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.189800 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f8f98d57b-66wrm" podStartSLOduration=3.9278178759999998 podStartE2EDuration="18.189782155s" podCreationTimestamp="2025-10-08 10:04:58 +0000 UTC" firstStartedPulling="2025-10-08 10:05:00.193214431 +0000 UTC m=+3195.440859660" lastFinishedPulling="2025-10-08 10:05:14.4551787 +0000 UTC m=+3209.702823939" observedRunningTime="2025-10-08 10:05:16.180068519 +0000 UTC m=+3211.427713768" watchObservedRunningTime="2025-10-08 10:05:16.189782155 +0000 UTC m=+3211.437427394" Oct 08 10:05:16 crc kubenswrapper[4744]: I1008 10:05:16.230299 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-66fcbb9755-w2nbw" podStartSLOduration=3.414033734 podStartE2EDuration="22.230283726s" podCreationTimestamp="2025-10-08 10:04:54 +0000 UTC" firstStartedPulling="2025-10-08 10:04:55.628770089 +0000 UTC m=+3190.876415328" lastFinishedPulling="2025-10-08 10:05:14.445020081 +0000 UTC m=+3209.692665320" observedRunningTime="2025-10-08 10:05:16.214835712 +0000 UTC m=+3211.462480961" watchObservedRunningTime="2025-10-08 10:05:16.230283726 +0000 UTC m=+3211.477928965" Oct 08 10:05:17 crc kubenswrapper[4744]: I1008 10:05:17.108537 4744 generic.go:334] "Generic (PLEG): container finished" podID="308c5a41-6d01-4684-9ab0-a23e0e372252" containerID="d494b0a7331293cdc351c41896e8f704de2faaf348611501ad220aa61e582341" exitCode=0 Oct 08 10:05:17 crc kubenswrapper[4744]: I1008 10:05:17.108610 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-5c7c-account-create-ww5bj" event={"ID":"308c5a41-6d01-4684-9ab0-a23e0e372252","Type":"ContainerDied","Data":"d494b0a7331293cdc351c41896e8f704de2faaf348611501ad220aa61e582341"} Oct 08 10:05:18 crc kubenswrapper[4744]: I1008 10:05:18.713285 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-5c7c-account-create-ww5bj" Oct 08 10:05:18 crc kubenswrapper[4744]: I1008 10:05:18.713331 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:05:18 crc kubenswrapper[4744]: I1008 10:05:18.713804 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:05:18 crc kubenswrapper[4744]: I1008 10:05:18.862269 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmt4s\" (UniqueName: \"kubernetes.io/projected/308c5a41-6d01-4684-9ab0-a23e0e372252-kube-api-access-tmt4s\") pod \"308c5a41-6d01-4684-9ab0-a23e0e372252\" (UID: \"308c5a41-6d01-4684-9ab0-a23e0e372252\") " Oct 08 10:05:18 crc kubenswrapper[4744]: I1008 10:05:18.894298 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308c5a41-6d01-4684-9ab0-a23e0e372252-kube-api-access-tmt4s" (OuterVolumeSpecName: "kube-api-access-tmt4s") pod "308c5a41-6d01-4684-9ab0-a23e0e372252" (UID: "308c5a41-6d01-4684-9ab0-a23e0e372252"). InnerVolumeSpecName "kube-api-access-tmt4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:05:18 crc kubenswrapper[4744]: I1008 10:05:18.964798 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmt4s\" (UniqueName: \"kubernetes.io/projected/308c5a41-6d01-4684-9ab0-a23e0e372252-kube-api-access-tmt4s\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:19 crc kubenswrapper[4744]: I1008 10:05:19.128489 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-5c7c-account-create-ww5bj" event={"ID":"308c5a41-6d01-4684-9ab0-a23e0e372252","Type":"ContainerDied","Data":"15480379b42e33d485eb87ce235ba093fbab8f9c51083fc60881170d4d9f3114"} Oct 08 10:05:19 crc kubenswrapper[4744]: I1008 10:05:19.128539 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15480379b42e33d485eb87ce235ba093fbab8f9c51083fc60881170d4d9f3114" Oct 08 10:05:19 crc kubenswrapper[4744]: I1008 10:05:19.128511 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-5c7c-account-create-ww5bj" Oct 08 10:05:19 crc kubenswrapper[4744]: I1008 10:05:19.141251 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 10:05:19 crc kubenswrapper[4744]: I1008 10:05:19.141486 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 10:05:19 crc kubenswrapper[4744]: I1008 10:05:19.152258 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:19 crc kubenswrapper[4744]: I1008 10:05:19.152419 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 10:05:19 crc kubenswrapper[4744]: I1008 10:05:19.158598 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 10:05:19 crc kubenswrapper[4744]: I1008 10:05:19.185386 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 10:05:19 crc kubenswrapper[4744]: I1008 10:05:19.385983 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:05:19 crc kubenswrapper[4744]: I1008 10:05:19.386965 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:05:24 crc kubenswrapper[4744]: I1008 10:05:24.670937 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.137699 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-s82nq"] Oct 08 10:05:25 crc kubenswrapper[4744]: E1008 10:05:25.138091 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="308c5a41-6d01-4684-9ab0-a23e0e372252" containerName="mariadb-account-create" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.138111 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="308c5a41-6d01-4684-9ab0-a23e0e372252" containerName="mariadb-account-create" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.138295 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="308c5a41-6d01-4684-9ab0-a23e0e372252" containerName="mariadb-account-create" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.138996 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.143940 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-jcgvd" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.144549 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.155483 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-s82nq"] Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.301660 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-config-data\") pod \"manila-db-sync-s82nq\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.301703 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-combined-ca-bundle\") pod \"manila-db-sync-s82nq\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.301776 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-job-config-data\") pod \"manila-db-sync-s82nq\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.301895 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6rfl\" (UniqueName: \"kubernetes.io/projected/975dbb77-5e8d-4192-885d-57910d229671-kube-api-access-p6rfl\") pod \"manila-db-sync-s82nq\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.403616 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6rfl\" (UniqueName: \"kubernetes.io/projected/975dbb77-5e8d-4192-885d-57910d229671-kube-api-access-p6rfl\") pod \"manila-db-sync-s82nq\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.404023 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-config-data\") pod \"manila-db-sync-s82nq\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.404043 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-combined-ca-bundle\") pod \"manila-db-sync-s82nq\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.404087 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-job-config-data\") pod \"manila-db-sync-s82nq\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.414034 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-combined-ca-bundle\") pod \"manila-db-sync-s82nq\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.415155 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-config-data\") pod \"manila-db-sync-s82nq\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.420982 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-job-config-data\") pod \"manila-db-sync-s82nq\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.430293 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6rfl\" (UniqueName: \"kubernetes.io/projected/975dbb77-5e8d-4192-885d-57910d229671-kube-api-access-p6rfl\") pod \"manila-db-sync-s82nq\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.444776 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:05:25 crc kubenswrapper[4744]: I1008 10:05:25.456134 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:26 crc kubenswrapper[4744]: I1008 10:05:26.215297 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-s82nq"] Oct 08 10:05:26 crc kubenswrapper[4744]: I1008 10:05:26.229737 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 10:05:27 crc kubenswrapper[4744]: I1008 10:05:27.229182 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-s82nq" event={"ID":"975dbb77-5e8d-4192-885d-57910d229671","Type":"ContainerStarted","Data":"c28a1b0a5b8cc95382f9dbec12a4c333eb7d921d348330aafc15ecb6be131a68"} Oct 08 10:05:28 crc kubenswrapper[4744]: I1008 10:05:28.718181 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f8f98d57b-66wrm" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Oct 08 10:05:29 crc kubenswrapper[4744]: I1008 10:05:29.387969 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5549798486-rbbgm" podUID="f337fa5f-1088-407d-a832-856ff504cd72" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.248:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.248:8443: connect: connection refused" Oct 08 10:05:29 crc kubenswrapper[4744]: I1008 10:05:29.457846 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:05:29 crc kubenswrapper[4744]: E1008 10:05:29.458361 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:05:33 crc kubenswrapper[4744]: I1008 10:05:33.308952 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-s82nq" event={"ID":"975dbb77-5e8d-4192-885d-57910d229671","Type":"ContainerStarted","Data":"18a5703374f620b27ffab52b4305c8616c73cc6d0743e3763c1d348b2656b366"} Oct 08 10:05:33 crc kubenswrapper[4744]: I1008 10:05:33.337922 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-s82nq" podStartSLOduration=2.632979869 podStartE2EDuration="8.337900495s" podCreationTimestamp="2025-10-08 10:05:25 +0000 UTC" firstStartedPulling="2025-10-08 10:05:26.229526846 +0000 UTC m=+3221.477172085" lastFinishedPulling="2025-10-08 10:05:31.934447472 +0000 UTC m=+3227.182092711" observedRunningTime="2025-10-08 10:05:33.330618566 +0000 UTC m=+3228.578263815" watchObservedRunningTime="2025-10-08 10:05:33.337900495 +0000 UTC m=+3228.585545734" Oct 08 10:05:38 crc kubenswrapper[4744]: I1008 10:05:38.714353 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f8f98d57b-66wrm" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Oct 08 10:05:39 crc kubenswrapper[4744]: I1008 10:05:39.386271 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5549798486-rbbgm" podUID="f337fa5f-1088-407d-a832-856ff504cd72" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.248:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.248:8443: connect: connection refused" Oct 08 10:05:43 crc kubenswrapper[4744]: I1008 10:05:43.433013 4744 generic.go:334] "Generic (PLEG): container finished" podID="975dbb77-5e8d-4192-885d-57910d229671" containerID="18a5703374f620b27ffab52b4305c8616c73cc6d0743e3763c1d348b2656b366" exitCode=0 Oct 08 10:05:43 crc kubenswrapper[4744]: I1008 10:05:43.433122 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-s82nq" event={"ID":"975dbb77-5e8d-4192-885d-57910d229671","Type":"ContainerDied","Data":"18a5703374f620b27ffab52b4305c8616c73cc6d0743e3763c1d348b2656b366"} Oct 08 10:05:43 crc kubenswrapper[4744]: I1008 10:05:43.456705 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:05:43 crc kubenswrapper[4744]: E1008 10:05:43.457029 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.043148 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.193342 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-job-config-data\") pod \"975dbb77-5e8d-4192-885d-57910d229671\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.193520 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-combined-ca-bundle\") pod \"975dbb77-5e8d-4192-885d-57910d229671\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.193546 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-config-data\") pod \"975dbb77-5e8d-4192-885d-57910d229671\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.193593 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6rfl\" (UniqueName: \"kubernetes.io/projected/975dbb77-5e8d-4192-885d-57910d229671-kube-api-access-p6rfl\") pod \"975dbb77-5e8d-4192-885d-57910d229671\" (UID: \"975dbb77-5e8d-4192-885d-57910d229671\") " Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.200598 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "975dbb77-5e8d-4192-885d-57910d229671" (UID: "975dbb77-5e8d-4192-885d-57910d229671"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.200619 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/975dbb77-5e8d-4192-885d-57910d229671-kube-api-access-p6rfl" (OuterVolumeSpecName: "kube-api-access-p6rfl") pod "975dbb77-5e8d-4192-885d-57910d229671" (UID: "975dbb77-5e8d-4192-885d-57910d229671"). InnerVolumeSpecName "kube-api-access-p6rfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.207196 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-config-data" (OuterVolumeSpecName: "config-data") pod "975dbb77-5e8d-4192-885d-57910d229671" (UID: "975dbb77-5e8d-4192-885d-57910d229671"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.243040 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "975dbb77-5e8d-4192-885d-57910d229671" (UID: "975dbb77-5e8d-4192-885d-57910d229671"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.295646 4744 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.295677 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.295685 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/975dbb77-5e8d-4192-885d-57910d229671-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.295694 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6rfl\" (UniqueName: \"kubernetes.io/projected/975dbb77-5e8d-4192-885d-57910d229671-kube-api-access-p6rfl\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.488449 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-s82nq" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.507909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-s82nq" event={"ID":"975dbb77-5e8d-4192-885d-57910d229671","Type":"ContainerDied","Data":"c28a1b0a5b8cc95382f9dbec12a4c333eb7d921d348330aafc15ecb6be131a68"} Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.507952 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c28a1b0a5b8cc95382f9dbec12a4c333eb7d921d348330aafc15ecb6be131a68" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.770501 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 08 10:05:45 crc kubenswrapper[4744]: E1008 10:05:45.771250 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="975dbb77-5e8d-4192-885d-57910d229671" containerName="manila-db-sync" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.771270 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="975dbb77-5e8d-4192-885d-57910d229671" containerName="manila-db-sync" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.771536 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="975dbb77-5e8d-4192-885d-57910d229671" containerName="manila-db-sync" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.772632 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.778674 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.784154 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.790130 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.790296 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-jcgvd" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.790865 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.802588 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.803868 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.820423 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.838347 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.916517 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-scripts\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.916565 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.916590 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.916615 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-ceph\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.916637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-config-data\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.917209 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88tlb\" (UniqueName: \"kubernetes.io/projected/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-kube-api-access-88tlb\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.917266 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.917294 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.917308 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-scripts\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.917328 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.917350 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f80f1604-c09c-4ee6-bb49-b867311bf1e6-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.917399 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-config-data\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.917423 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.917442 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grxq2\" (UniqueName: \"kubernetes.io/projected/f80f1604-c09c-4ee6-bb49-b867311bf1e6-kube-api-access-grxq2\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.944483 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7887c4559f-jqcwg"] Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.946306 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:45 crc kubenswrapper[4744]: I1008 10:05:45.971107 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7887c4559f-jqcwg"] Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.018954 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.019894 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.019933 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-scripts\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.019962 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.019988 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f80f1604-c09c-4ee6-bb49-b867311bf1e6-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.020033 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-config-data\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.020060 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.020088 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grxq2\" (UniqueName: \"kubernetes.io/projected/f80f1604-c09c-4ee6-bb49-b867311bf1e6-kube-api-access-grxq2\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.020175 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-scripts\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.020198 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.020222 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.020246 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-ceph\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.020267 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-config-data\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.020289 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88tlb\" (UniqueName: \"kubernetes.io/projected/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-kube-api-access-88tlb\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.023546 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f80f1604-c09c-4ee6-bb49-b867311bf1e6-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.024858 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.026198 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.027426 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.027496 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.030113 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-config-data\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.030654 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-scripts\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.042910 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grxq2\" (UniqueName: \"kubernetes.io/projected/f80f1604-c09c-4ee6-bb49-b867311bf1e6-kube-api-access-grxq2\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.043877 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-scripts\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.046902 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88tlb\" (UniqueName: \"kubernetes.io/projected/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-kube-api-access-88tlb\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.049623 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.049999 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-ceph\") pod \"manila-share-share1-0\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.061168 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.063102 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-config-data\") pod \"manila-scheduler-0\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.096792 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.113597 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.114046 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.115238 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121219 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-config-data\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121269 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121294 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45bkk\" (UniqueName: \"kubernetes.io/projected/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-kube-api-access-45bkk\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121325 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-ovsdbserver-nb\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121361 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-openstack-edpm-ipam\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121399 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h8vl\" (UniqueName: \"kubernetes.io/projected/0d390563-4f19-4b41-abdd-b3f449a45ef0-kube-api-access-5h8vl\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121438 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-dns-svc\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121464 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-config-data-custom\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121492 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-scripts\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121512 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-config\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121540 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d390563-4f19-4b41-abdd-b3f449a45ef0-etc-machine-id\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121564 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-ovsdbserver-sb\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.121581 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d390563-4f19-4b41-abdd-b3f449a45ef0-logs\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.134597 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.161559 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.223863 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-ovsdbserver-sb\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.223914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d390563-4f19-4b41-abdd-b3f449a45ef0-logs\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.223944 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-config-data\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.223973 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.223993 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45bkk\" (UniqueName: \"kubernetes.io/projected/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-kube-api-access-45bkk\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.224027 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-ovsdbserver-nb\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.224065 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-openstack-edpm-ipam\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.224088 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h8vl\" (UniqueName: \"kubernetes.io/projected/0d390563-4f19-4b41-abdd-b3f449a45ef0-kube-api-access-5h8vl\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.225825 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-ovsdbserver-sb\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.225982 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-dns-svc\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.226025 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-config-data-custom\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.226081 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-scripts\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.226106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-config\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.226147 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d390563-4f19-4b41-abdd-b3f449a45ef0-etc-machine-id\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.226215 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d390563-4f19-4b41-abdd-b3f449a45ef0-etc-machine-id\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.229121 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-dns-svc\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.229686 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-ovsdbserver-nb\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.230559 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d390563-4f19-4b41-abdd-b3f449a45ef0-logs\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.232619 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-config\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.237018 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-config-data-custom\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.241951 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-openstack-edpm-ipam\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.242744 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-config-data\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.245283 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45bkk\" (UniqueName: \"kubernetes.io/projected/084ffd95-c3f3-4340-bc26-d3c2c427c8ad-kube-api-access-45bkk\") pod \"dnsmasq-dns-7887c4559f-jqcwg\" (UID: \"084ffd95-c3f3-4340-bc26-d3c2c427c8ad\") " pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.246911 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h8vl\" (UniqueName: \"kubernetes.io/projected/0d390563-4f19-4b41-abdd-b3f449a45ef0-kube-api-access-5h8vl\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.247539 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.249646 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-scripts\") pod \"manila-api-0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.266852 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.400346 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.545078 4744 generic.go:334] "Generic (PLEG): container finished" podID="8b41d1ed-6119-42a5-8ee7-d4e820d595a8" containerID="bf53ea7e94ab1a9dc1f1252d25d0883df0ff4a2fae485a02e28790d7e8156bff" exitCode=137 Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.545105 4744 generic.go:334] "Generic (PLEG): container finished" podID="8b41d1ed-6119-42a5-8ee7-d4e820d595a8" containerID="411b8a7ef38838dd31b1a9be5935619da8e930b2e99f4869d6776712e32d926c" exitCode=137 Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.545142 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f866cb5ff-dkrbz" event={"ID":"8b41d1ed-6119-42a5-8ee7-d4e820d595a8","Type":"ContainerDied","Data":"bf53ea7e94ab1a9dc1f1252d25d0883df0ff4a2fae485a02e28790d7e8156bff"} Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.545168 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f866cb5ff-dkrbz" event={"ID":"8b41d1ed-6119-42a5-8ee7-d4e820d595a8","Type":"ContainerDied","Data":"411b8a7ef38838dd31b1a9be5935619da8e930b2e99f4869d6776712e32d926c"} Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.549912 4744 generic.go:334] "Generic (PLEG): container finished" podID="b492a22d-99c3-4ec6-8ab6-be3875679018" containerID="6961c20b15cbd567c5c14acb0c8c3f875a72fb0094f0bdf4fa1888723670f461" exitCode=137 Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.549932 4744 generic.go:334] "Generic (PLEG): container finished" podID="b492a22d-99c3-4ec6-8ab6-be3875679018" containerID="9d230f0533ead7fc718808c57ac2da9650713e9e95833d6f29715c32d1a60b99" exitCode=137 Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.549952 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66fcbb9755-w2nbw" event={"ID":"b492a22d-99c3-4ec6-8ab6-be3875679018","Type":"ContainerDied","Data":"6961c20b15cbd567c5c14acb0c8c3f875a72fb0094f0bdf4fa1888723670f461"} Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.549974 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66fcbb9755-w2nbw" event={"ID":"b492a22d-99c3-4ec6-8ab6-be3875679018","Type":"ContainerDied","Data":"9d230f0533ead7fc718808c57ac2da9650713e9e95833d6f29715c32d1a60b99"} Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.933123 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.951491 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl6sz\" (UniqueName: \"kubernetes.io/projected/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-kube-api-access-vl6sz\") pod \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.951542 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-logs\") pod \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.951626 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-horizon-secret-key\") pod \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.951729 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-config-data\") pod \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.951807 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-scripts\") pod \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\" (UID: \"8b41d1ed-6119-42a5-8ee7-d4e820d595a8\") " Oct 08 10:05:46 crc kubenswrapper[4744]: I1008 10:05:46.954037 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-logs" (OuterVolumeSpecName: "logs") pod "8b41d1ed-6119-42a5-8ee7-d4e820d595a8" (UID: "8b41d1ed-6119-42a5-8ee7-d4e820d595a8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.054752 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-logs\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.066448 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8b41d1ed-6119-42a5-8ee7-d4e820d595a8" (UID: "8b41d1ed-6119-42a5-8ee7-d4e820d595a8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.066542 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-scripts" (OuterVolumeSpecName: "scripts") pod "8b41d1ed-6119-42a5-8ee7-d4e820d595a8" (UID: "8b41d1ed-6119-42a5-8ee7-d4e820d595a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.069201 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-kube-api-access-vl6sz" (OuterVolumeSpecName: "kube-api-access-vl6sz") pod "8b41d1ed-6119-42a5-8ee7-d4e820d595a8" (UID: "8b41d1ed-6119-42a5-8ee7-d4e820d595a8"). InnerVolumeSpecName "kube-api-access-vl6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.084205 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-config-data" (OuterVolumeSpecName: "config-data") pod "8b41d1ed-6119-42a5-8ee7-d4e820d595a8" (UID: "8b41d1ed-6119-42a5-8ee7-d4e820d595a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.134099 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.156293 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl6sz\" (UniqueName: \"kubernetes.io/projected/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-kube-api-access-vl6sz\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.156330 4744 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.156339 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.156348 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b41d1ed-6119-42a5-8ee7-d4e820d595a8-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.203831 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.257579 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b492a22d-99c3-4ec6-8ab6-be3875679018-logs\") pod \"b492a22d-99c3-4ec6-8ab6-be3875679018\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.257883 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b492a22d-99c3-4ec6-8ab6-be3875679018-config-data\") pod \"b492a22d-99c3-4ec6-8ab6-be3875679018\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.258004 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcd8v\" (UniqueName: \"kubernetes.io/projected/b492a22d-99c3-4ec6-8ab6-be3875679018-kube-api-access-wcd8v\") pod \"b492a22d-99c3-4ec6-8ab6-be3875679018\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.258083 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b492a22d-99c3-4ec6-8ab6-be3875679018-horizon-secret-key\") pod \"b492a22d-99c3-4ec6-8ab6-be3875679018\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.258162 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b492a22d-99c3-4ec6-8ab6-be3875679018-scripts\") pod \"b492a22d-99c3-4ec6-8ab6-be3875679018\" (UID: \"b492a22d-99c3-4ec6-8ab6-be3875679018\") " Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.261889 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b492a22d-99c3-4ec6-8ab6-be3875679018-logs" (OuterVolumeSpecName: "logs") pod "b492a22d-99c3-4ec6-8ab6-be3875679018" (UID: "b492a22d-99c3-4ec6-8ab6-be3875679018"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.264039 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b492a22d-99c3-4ec6-8ab6-be3875679018-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b492a22d-99c3-4ec6-8ab6-be3875679018" (UID: "b492a22d-99c3-4ec6-8ab6-be3875679018"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.353703 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b492a22d-99c3-4ec6-8ab6-be3875679018-config-data" (OuterVolumeSpecName: "config-data") pod "b492a22d-99c3-4ec6-8ab6-be3875679018" (UID: "b492a22d-99c3-4ec6-8ab6-be3875679018"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:05:47 crc kubenswrapper[4744]: W1008 10:05:47.363885 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod084ffd95_c3f3_4340_bc26_d3c2c427c8ad.slice/crio-ba89ba298ec22725b616662e1a248fc170467e97f7d572576192dc2c395f9da4 WatchSource:0}: Error finding container ba89ba298ec22725b616662e1a248fc170467e97f7d572576192dc2c395f9da4: Status 404 returned error can't find the container with id ba89ba298ec22725b616662e1a248fc170467e97f7d572576192dc2c395f9da4 Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.375258 4744 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b492a22d-99c3-4ec6-8ab6-be3875679018-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.375283 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b492a22d-99c3-4ec6-8ab6-be3875679018-logs\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.375299 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b492a22d-99c3-4ec6-8ab6-be3875679018-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.376296 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b492a22d-99c3-4ec6-8ab6-be3875679018-kube-api-access-wcd8v" (OuterVolumeSpecName: "kube-api-access-wcd8v") pod "b492a22d-99c3-4ec6-8ab6-be3875679018" (UID: "b492a22d-99c3-4ec6-8ab6-be3875679018"). InnerVolumeSpecName "kube-api-access-wcd8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.413295 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7887c4559f-jqcwg"] Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.472652 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b492a22d-99c3-4ec6-8ab6-be3875679018-scripts" (OuterVolumeSpecName: "scripts") pod "b492a22d-99c3-4ec6-8ab6-be3875679018" (UID: "b492a22d-99c3-4ec6-8ab6-be3875679018"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.480174 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcd8v\" (UniqueName: \"kubernetes.io/projected/b492a22d-99c3-4ec6-8ab6-be3875679018-kube-api-access-wcd8v\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.480198 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b492a22d-99c3-4ec6-8ab6-be3875679018-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.533456 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.578842 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f80f1604-c09c-4ee6-bb49-b867311bf1e6","Type":"ContainerStarted","Data":"f09d45847727726e2e322b87cd4717a99def8f9ecbd0c04acb83b161dbaf8760"} Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.587285 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f866cb5ff-dkrbz" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.587282 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f866cb5ff-dkrbz" event={"ID":"8b41d1ed-6119-42a5-8ee7-d4e820d595a8","Type":"ContainerDied","Data":"6a7604bd17f7c309ce0d185ad6662276f2cd1383d3ecfc3d6977189019080e68"} Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.587559 4744 scope.go:117] "RemoveContainer" containerID="bf53ea7e94ab1a9dc1f1252d25d0883df0ff4a2fae485a02e28790d7e8156bff" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.591181 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" event={"ID":"084ffd95-c3f3-4340-bc26-d3c2c427c8ad","Type":"ContainerStarted","Data":"ba89ba298ec22725b616662e1a248fc170467e97f7d572576192dc2c395f9da4"} Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.594790 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66fcbb9755-w2nbw" event={"ID":"b492a22d-99c3-4ec6-8ab6-be3875679018","Type":"ContainerDied","Data":"fd5ab7654834adf9e2c03431cb47aee23483bb16c06f10ce19ea1c78fdadddf7"} Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.594913 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66fcbb9755-w2nbw" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.601852 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c","Type":"ContainerStarted","Data":"1dac840eabf2adfa1b9b91877b83b6eb2446c4b25bc8627eed40ddf4f49fd4e4"} Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.647203 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f866cb5ff-dkrbz"] Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.662139 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5f866cb5ff-dkrbz"] Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.673547 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66fcbb9755-w2nbw"] Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.686236 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-66fcbb9755-w2nbw"] Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.864815 4744 scope.go:117] "RemoveContainer" containerID="411b8a7ef38838dd31b1a9be5935619da8e930b2e99f4869d6776712e32d926c" Oct 08 10:05:47 crc kubenswrapper[4744]: I1008 10:05:47.986210 4744 scope.go:117] "RemoveContainer" containerID="6961c20b15cbd567c5c14acb0c8c3f875a72fb0094f0bdf4fa1888723670f461" Oct 08 10:05:48 crc kubenswrapper[4744]: I1008 10:05:48.165390 4744 scope.go:117] "RemoveContainer" containerID="9d230f0533ead7fc718808c57ac2da9650713e9e95833d6f29715c32d1a60b99" Oct 08 10:05:48 crc kubenswrapper[4744]: W1008 10:05:48.341566 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d390563_4f19_4b41_abdd_b3f449a45ef0.slice/crio-54a46b352e87a86386e3e898fd69f601578ab7d2b5a9e00141b25122596b9ceb WatchSource:0}: Error finding container 54a46b352e87a86386e3e898fd69f601578ab7d2b5a9e00141b25122596b9ceb: Status 404 returned error can't find the container with id 54a46b352e87a86386e3e898fd69f601578ab7d2b5a9e00141b25122596b9ceb Oct 08 10:05:48 crc kubenswrapper[4744]: I1008 10:05:48.352935 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 08 10:05:48 crc kubenswrapper[4744]: I1008 10:05:48.648850 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0d390563-4f19-4b41-abdd-b3f449a45ef0","Type":"ContainerStarted","Data":"54a46b352e87a86386e3e898fd69f601578ab7d2b5a9e00141b25122596b9ceb"} Oct 08 10:05:48 crc kubenswrapper[4744]: I1008 10:05:48.656091 4744 generic.go:334] "Generic (PLEG): container finished" podID="084ffd95-c3f3-4340-bc26-d3c2c427c8ad" containerID="891d611b8f2c505542e0e6df070ab57d51a8a135e76c610e5d792ad247dc7aee" exitCode=0 Oct 08 10:05:48 crc kubenswrapper[4744]: I1008 10:05:48.656153 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" event={"ID":"084ffd95-c3f3-4340-bc26-d3c2c427c8ad","Type":"ContainerDied","Data":"891d611b8f2c505542e0e6df070ab57d51a8a135e76c610e5d792ad247dc7aee"} Oct 08 10:05:48 crc kubenswrapper[4744]: I1008 10:05:48.674615 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 08 10:05:49 crc kubenswrapper[4744]: I1008 10:05:49.522757 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b41d1ed-6119-42a5-8ee7-d4e820d595a8" path="/var/lib/kubelet/pods/8b41d1ed-6119-42a5-8ee7-d4e820d595a8/volumes" Oct 08 10:05:49 crc kubenswrapper[4744]: I1008 10:05:49.527488 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b492a22d-99c3-4ec6-8ab6-be3875679018" path="/var/lib/kubelet/pods/b492a22d-99c3-4ec6-8ab6-be3875679018/volumes" Oct 08 10:05:49 crc kubenswrapper[4744]: I1008 10:05:49.743716 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0d390563-4f19-4b41-abdd-b3f449a45ef0","Type":"ContainerStarted","Data":"0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45"} Oct 08 10:05:49 crc kubenswrapper[4744]: I1008 10:05:49.808903 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" event={"ID":"084ffd95-c3f3-4340-bc26-d3c2c427c8ad","Type":"ContainerStarted","Data":"3f6b663b314e76e961c4eb922bc48eb940bbe8942b4bf4eb9ffb3412aed6063a"} Oct 08 10:05:49 crc kubenswrapper[4744]: I1008 10:05:49.808983 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:49 crc kubenswrapper[4744]: I1008 10:05:49.819118 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f80f1604-c09c-4ee6-bb49-b867311bf1e6","Type":"ContainerStarted","Data":"e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3"} Oct 08 10:05:49 crc kubenswrapper[4744]: I1008 10:05:49.847600 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" podStartSLOduration=4.84758211 podStartE2EDuration="4.84758211s" podCreationTimestamp="2025-10-08 10:05:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:05:49.834892362 +0000 UTC m=+3245.082537611" watchObservedRunningTime="2025-10-08 10:05:49.84758211 +0000 UTC m=+3245.095227349" Oct 08 10:05:50 crc kubenswrapper[4744]: I1008 10:05:50.829210 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0d390563-4f19-4b41-abdd-b3f449a45ef0","Type":"ContainerStarted","Data":"268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc"} Oct 08 10:05:50 crc kubenswrapper[4744]: I1008 10:05:50.829718 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 08 10:05:50 crc kubenswrapper[4744]: I1008 10:05:50.829617 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="0d390563-4f19-4b41-abdd-b3f449a45ef0" containerName="manila-api" containerID="cri-o://268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc" gracePeriod=30 Oct 08 10:05:50 crc kubenswrapper[4744]: I1008 10:05:50.829320 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="0d390563-4f19-4b41-abdd-b3f449a45ef0" containerName="manila-api-log" containerID="cri-o://0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45" gracePeriod=30 Oct 08 10:05:50 crc kubenswrapper[4744]: I1008 10:05:50.841523 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f80f1604-c09c-4ee6-bb49-b867311bf1e6","Type":"ContainerStarted","Data":"231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6"} Oct 08 10:05:50 crc kubenswrapper[4744]: I1008 10:05:50.876072 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=4.876043142 podStartE2EDuration="4.876043142s" podCreationTimestamp="2025-10-08 10:05:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:05:50.85517856 +0000 UTC m=+3246.102823799" watchObservedRunningTime="2025-10-08 10:05:50.876043142 +0000 UTC m=+3246.123688381" Oct 08 10:05:50 crc kubenswrapper[4744]: I1008 10:05:50.880787 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=4.7866768010000005 podStartE2EDuration="5.880774622s" podCreationTimestamp="2025-10-08 10:05:45 +0000 UTC" firstStartedPulling="2025-10-08 10:05:47.153554427 +0000 UTC m=+3242.401199666" lastFinishedPulling="2025-10-08 10:05:48.247652248 +0000 UTC m=+3243.495297487" observedRunningTime="2025-10-08 10:05:50.873482501 +0000 UTC m=+3246.121127740" watchObservedRunningTime="2025-10-08 10:05:50.880774622 +0000 UTC m=+3246.128419891" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.863301 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.866758 4744 generic.go:334] "Generic (PLEG): container finished" podID="0d390563-4f19-4b41-abdd-b3f449a45ef0" containerID="268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc" exitCode=0 Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.866799 4744 generic.go:334] "Generic (PLEG): container finished" podID="0d390563-4f19-4b41-abdd-b3f449a45ef0" containerID="0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45" exitCode=143 Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.867171 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0d390563-4f19-4b41-abdd-b3f449a45ef0","Type":"ContainerDied","Data":"268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc"} Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.867237 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0d390563-4f19-4b41-abdd-b3f449a45ef0","Type":"ContainerDied","Data":"0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45"} Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.867251 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0d390563-4f19-4b41-abdd-b3f449a45ef0","Type":"ContainerDied","Data":"54a46b352e87a86386e3e898fd69f601578ab7d2b5a9e00141b25122596b9ceb"} Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.867277 4744 scope.go:117] "RemoveContainer" containerID="268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.885495 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-config-data\") pod \"0d390563-4f19-4b41-abdd-b3f449a45ef0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.885569 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d390563-4f19-4b41-abdd-b3f449a45ef0-logs\") pod \"0d390563-4f19-4b41-abdd-b3f449a45ef0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.885621 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h8vl\" (UniqueName: \"kubernetes.io/projected/0d390563-4f19-4b41-abdd-b3f449a45ef0-kube-api-access-5h8vl\") pod \"0d390563-4f19-4b41-abdd-b3f449a45ef0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.885680 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-scripts\") pod \"0d390563-4f19-4b41-abdd-b3f449a45ef0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.885707 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-combined-ca-bundle\") pod \"0d390563-4f19-4b41-abdd-b3f449a45ef0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.885820 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-config-data-custom\") pod \"0d390563-4f19-4b41-abdd-b3f449a45ef0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.886011 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d390563-4f19-4b41-abdd-b3f449a45ef0-etc-machine-id\") pod \"0d390563-4f19-4b41-abdd-b3f449a45ef0\" (UID: \"0d390563-4f19-4b41-abdd-b3f449a45ef0\") " Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.887900 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d390563-4f19-4b41-abdd-b3f449a45ef0-logs" (OuterVolumeSpecName: "logs") pod "0d390563-4f19-4b41-abdd-b3f449a45ef0" (UID: "0d390563-4f19-4b41-abdd-b3f449a45ef0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.888462 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d390563-4f19-4b41-abdd-b3f449a45ef0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0d390563-4f19-4b41-abdd-b3f449a45ef0" (UID: "0d390563-4f19-4b41-abdd-b3f449a45ef0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.906502 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d390563-4f19-4b41-abdd-b3f449a45ef0-kube-api-access-5h8vl" (OuterVolumeSpecName: "kube-api-access-5h8vl") pod "0d390563-4f19-4b41-abdd-b3f449a45ef0" (UID: "0d390563-4f19-4b41-abdd-b3f449a45ef0"). InnerVolumeSpecName "kube-api-access-5h8vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.906643 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0d390563-4f19-4b41-abdd-b3f449a45ef0" (UID: "0d390563-4f19-4b41-abdd-b3f449a45ef0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.916571 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-scripts" (OuterVolumeSpecName: "scripts") pod "0d390563-4f19-4b41-abdd-b3f449a45ef0" (UID: "0d390563-4f19-4b41-abdd-b3f449a45ef0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.962917 4744 scope.go:117] "RemoveContainer" containerID="0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.990208 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h8vl\" (UniqueName: \"kubernetes.io/projected/0d390563-4f19-4b41-abdd-b3f449a45ef0-kube-api-access-5h8vl\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.990253 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.990264 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.990275 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0d390563-4f19-4b41-abdd-b3f449a45ef0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.990287 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d390563-4f19-4b41-abdd-b3f449a45ef0-logs\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:51 crc kubenswrapper[4744]: I1008 10:05:51.997707 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d390563-4f19-4b41-abdd-b3f449a45ef0" (UID: "0d390563-4f19-4b41-abdd-b3f449a45ef0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.019814 4744 scope.go:117] "RemoveContainer" containerID="268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc" Oct 08 10:05:52 crc kubenswrapper[4744]: E1008 10:05:52.020283 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc\": container with ID starting with 268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc not found: ID does not exist" containerID="268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.020405 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc"} err="failed to get container status \"268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc\": rpc error: code = NotFound desc = could not find container \"268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc\": container with ID starting with 268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc not found: ID does not exist" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.020479 4744 scope.go:117] "RemoveContainer" containerID="0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45" Oct 08 10:05:52 crc kubenswrapper[4744]: E1008 10:05:52.020759 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45\": container with ID starting with 0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45 not found: ID does not exist" containerID="0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.020867 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45"} err="failed to get container status \"0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45\": rpc error: code = NotFound desc = could not find container \"0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45\": container with ID starting with 0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45 not found: ID does not exist" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.020971 4744 scope.go:117] "RemoveContainer" containerID="268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.021235 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc"} err="failed to get container status \"268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc\": rpc error: code = NotFound desc = could not find container \"268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc\": container with ID starting with 268b016bb301c2bf4b6fa968c633336082a65fb439bc279c833027fd364d4dfc not found: ID does not exist" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.021316 4744 scope.go:117] "RemoveContainer" containerID="0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.022857 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45"} err="failed to get container status \"0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45\": rpc error: code = NotFound desc = could not find container \"0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45\": container with ID starting with 0f19740522d3a2a418d955ace2af428108933f0a03ddf61d5cf4d8780e2e5a45 not found: ID does not exist" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.057490 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-config-data" (OuterVolumeSpecName: "config-data") pod "0d390563-4f19-4b41-abdd-b3f449a45ef0" (UID: "0d390563-4f19-4b41-abdd-b3f449a45ef0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.091299 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.091519 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d390563-4f19-4b41-abdd-b3f449a45ef0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.739287 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.740965 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.879635 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.920143 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 08 10:05:52 crc kubenswrapper[4744]: I1008 10:05:52.935456 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.065883 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 08 10:05:53 crc kubenswrapper[4744]: E1008 10:05:53.066982 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b41d1ed-6119-42a5-8ee7-d4e820d595a8" containerName="horizon-log" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.067001 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b41d1ed-6119-42a5-8ee7-d4e820d595a8" containerName="horizon-log" Oct 08 10:05:53 crc kubenswrapper[4744]: E1008 10:05:53.067024 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b492a22d-99c3-4ec6-8ab6-be3875679018" containerName="horizon-log" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.067031 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b492a22d-99c3-4ec6-8ab6-be3875679018" containerName="horizon-log" Oct 08 10:05:53 crc kubenswrapper[4744]: E1008 10:05:53.067048 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d390563-4f19-4b41-abdd-b3f449a45ef0" containerName="manila-api-log" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.067055 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d390563-4f19-4b41-abdd-b3f449a45ef0" containerName="manila-api-log" Oct 08 10:05:53 crc kubenswrapper[4744]: E1008 10:05:53.067096 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b492a22d-99c3-4ec6-8ab6-be3875679018" containerName="horizon" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.067104 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b492a22d-99c3-4ec6-8ab6-be3875679018" containerName="horizon" Oct 08 10:05:53 crc kubenswrapper[4744]: E1008 10:05:53.067121 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d390563-4f19-4b41-abdd-b3f449a45ef0" containerName="manila-api" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.067127 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d390563-4f19-4b41-abdd-b3f449a45ef0" containerName="manila-api" Oct 08 10:05:53 crc kubenswrapper[4744]: E1008 10:05:53.067145 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b41d1ed-6119-42a5-8ee7-d4e820d595a8" containerName="horizon" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.067153 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b41d1ed-6119-42a5-8ee7-d4e820d595a8" containerName="horizon" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.067669 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b492a22d-99c3-4ec6-8ab6-be3875679018" containerName="horizon-log" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.067693 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b41d1ed-6119-42a5-8ee7-d4e820d595a8" containerName="horizon" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.067705 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b492a22d-99c3-4ec6-8ab6-be3875679018" containerName="horizon" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.067722 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b41d1ed-6119-42a5-8ee7-d4e820d595a8" containerName="horizon-log" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.067739 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d390563-4f19-4b41-abdd-b3f449a45ef0" containerName="manila-api" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.067757 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d390563-4f19-4b41-abdd-b3f449a45ef0" containerName="manila-api-log" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.073723 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.076062 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.094413 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.095517 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.097675 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.228194 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1003e3fc-33a3-45d1-a1b7-9f383af6e336-logs\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.228240 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhd79\" (UniqueName: \"kubernetes.io/projected/1003e3fc-33a3-45d1-a1b7-9f383af6e336-kube-api-access-jhd79\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.228281 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-internal-tls-certs\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.228323 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-scripts\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.228347 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-public-tls-certs\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.228400 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-config-data\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.228426 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-config-data-custom\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.228454 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1003e3fc-33a3-45d1-a1b7-9f383af6e336-etc-machine-id\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.228471 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.331555 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-config-data-custom\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.331634 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1003e3fc-33a3-45d1-a1b7-9f383af6e336-etc-machine-id\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.331668 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.331765 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1003e3fc-33a3-45d1-a1b7-9f383af6e336-logs\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.331801 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhd79\" (UniqueName: \"kubernetes.io/projected/1003e3fc-33a3-45d1-a1b7-9f383af6e336-kube-api-access-jhd79\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.331845 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-internal-tls-certs\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.331888 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-scripts\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.331913 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-public-tls-certs\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.331953 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-config-data\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.332287 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1003e3fc-33a3-45d1-a1b7-9f383af6e336-logs\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.331762 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1003e3fc-33a3-45d1-a1b7-9f383af6e336-etc-machine-id\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.338456 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-scripts\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.340360 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.341524 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-config-data\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.343187 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-config-data-custom\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.355317 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-public-tls-certs\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.375576 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhd79\" (UniqueName: \"kubernetes.io/projected/1003e3fc-33a3-45d1-a1b7-9f383af6e336-kube-api-access-jhd79\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.375588 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1003e3fc-33a3-45d1-a1b7-9f383af6e336-internal-tls-certs\") pod \"manila-api-0\" (UID: \"1003e3fc-33a3-45d1-a1b7-9f383af6e336\") " pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.415036 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 08 10:05:53 crc kubenswrapper[4744]: I1008 10:05:53.481527 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d390563-4f19-4b41-abdd-b3f449a45ef0" path="/var/lib/kubelet/pods/0d390563-4f19-4b41-abdd-b3f449a45ef0/volumes" Oct 08 10:05:54 crc kubenswrapper[4744]: I1008 10:05:54.116523 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 08 10:05:54 crc kubenswrapper[4744]: I1008 10:05:54.924047 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"1003e3fc-33a3-45d1-a1b7-9f383af6e336","Type":"ContainerStarted","Data":"d09db3175ba2e17dc4a86eded324776312f012f1f2cfcfb21d7a797f612bf90b"} Oct 08 10:05:55 crc kubenswrapper[4744]: I1008 10:05:55.074925 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5549798486-rbbgm" Oct 08 10:05:55 crc kubenswrapper[4744]: I1008 10:05:55.151322 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f8f98d57b-66wrm"] Oct 08 10:05:55 crc kubenswrapper[4744]: I1008 10:05:55.151573 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f8f98d57b-66wrm" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon-log" containerID="cri-o://621e50ea58c51b02bda6c75857ac031b6d770e3776c4efe36a0937f13140b393" gracePeriod=30 Oct 08 10:05:55 crc kubenswrapper[4744]: I1008 10:05:55.152003 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f8f98d57b-66wrm" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon" containerID="cri-o://2139c1999ee135e2b62d6e294e9ec99ea5173c7f974b9cfaa2786d28f6dac0bb" gracePeriod=30 Oct 08 10:05:55 crc kubenswrapper[4744]: I1008 10:05:55.170250 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7f8f98d57b-66wrm" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Oct 08 10:05:56 crc kubenswrapper[4744]: I1008 10:05:56.097928 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 08 10:05:56 crc kubenswrapper[4744]: I1008 10:05:56.269541 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7887c4559f-jqcwg" Oct 08 10:05:56 crc kubenswrapper[4744]: I1008 10:05:56.435867 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-867c8fd5c5-wkswc"] Oct 08 10:05:56 crc kubenswrapper[4744]: I1008 10:05:56.436658 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" podUID="65ab44b6-3478-42d3-b9cb-64f911f13342" containerName="dnsmasq-dns" containerID="cri-o://754ac2af91ca329df2d3d3340340cb5a2294499e668c5b00a0374f7e45b216dc" gracePeriod=10 Oct 08 10:05:56 crc kubenswrapper[4744]: I1008 10:05:56.946851 4744 generic.go:334] "Generic (PLEG): container finished" podID="65ab44b6-3478-42d3-b9cb-64f911f13342" containerID="754ac2af91ca329df2d3d3340340cb5a2294499e668c5b00a0374f7e45b216dc" exitCode=0 Oct 08 10:05:56 crc kubenswrapper[4744]: I1008 10:05:56.946898 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" event={"ID":"65ab44b6-3478-42d3-b9cb-64f911f13342","Type":"ContainerDied","Data":"754ac2af91ca329df2d3d3340340cb5a2294499e668c5b00a0374f7e45b216dc"} Oct 08 10:05:57 crc kubenswrapper[4744]: I1008 10:05:57.453757 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:05:57 crc kubenswrapper[4744]: E1008 10:05:57.454111 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.393284 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.468272 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-config\") pod \"65ab44b6-3478-42d3-b9cb-64f911f13342\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.468353 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-openstack-edpm-ipam\") pod \"65ab44b6-3478-42d3-b9cb-64f911f13342\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.468393 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-dns-svc\") pod \"65ab44b6-3478-42d3-b9cb-64f911f13342\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.468653 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-ovsdbserver-nb\") pod \"65ab44b6-3478-42d3-b9cb-64f911f13342\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.468730 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw822\" (UniqueName: \"kubernetes.io/projected/65ab44b6-3478-42d3-b9cb-64f911f13342-kube-api-access-dw822\") pod \"65ab44b6-3478-42d3-b9cb-64f911f13342\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.468790 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-ovsdbserver-sb\") pod \"65ab44b6-3478-42d3-b9cb-64f911f13342\" (UID: \"65ab44b6-3478-42d3-b9cb-64f911f13342\") " Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.491466 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ab44b6-3478-42d3-b9cb-64f911f13342-kube-api-access-dw822" (OuterVolumeSpecName: "kube-api-access-dw822") pod "65ab44b6-3478-42d3-b9cb-64f911f13342" (UID: "65ab44b6-3478-42d3-b9cb-64f911f13342"). InnerVolumeSpecName "kube-api-access-dw822". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.571842 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw822\" (UniqueName: \"kubernetes.io/projected/65ab44b6-3478-42d3-b9cb-64f911f13342-kube-api-access-dw822\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.574733 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7f8f98d57b-66wrm" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:50568->10.217.0.247:8443: read: connection reset by peer" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.600875 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "65ab44b6-3478-42d3-b9cb-64f911f13342" (UID: "65ab44b6-3478-42d3-b9cb-64f911f13342"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.607594 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "65ab44b6-3478-42d3-b9cb-64f911f13342" (UID: "65ab44b6-3478-42d3-b9cb-64f911f13342"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.621210 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-config" (OuterVolumeSpecName: "config") pod "65ab44b6-3478-42d3-b9cb-64f911f13342" (UID: "65ab44b6-3478-42d3-b9cb-64f911f13342"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.623577 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "65ab44b6-3478-42d3-b9cb-64f911f13342" (UID: "65ab44b6-3478-42d3-b9cb-64f911f13342"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.643737 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "65ab44b6-3478-42d3-b9cb-64f911f13342" (UID: "65ab44b6-3478-42d3-b9cb-64f911f13342"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.673547 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.673580 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-config\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.673590 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.673598 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.673607 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/65ab44b6-3478-42d3-b9cb-64f911f13342-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 10:05:58 crc kubenswrapper[4744]: I1008 10:05:58.714341 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7f8f98d57b-66wrm" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Oct 08 10:05:59 crc kubenswrapper[4744]: I1008 10:05:59.002144 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c","Type":"ContainerStarted","Data":"f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd"} Oct 08 10:05:59 crc kubenswrapper[4744]: I1008 10:05:59.053012 4744 generic.go:334] "Generic (PLEG): container finished" podID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerID="2139c1999ee135e2b62d6e294e9ec99ea5173c7f974b9cfaa2786d28f6dac0bb" exitCode=0 Oct 08 10:05:59 crc kubenswrapper[4744]: I1008 10:05:59.053149 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f8f98d57b-66wrm" event={"ID":"49150e8b-4e4b-467f-ae7b-a6d14b63554d","Type":"ContainerDied","Data":"2139c1999ee135e2b62d6e294e9ec99ea5173c7f974b9cfaa2786d28f6dac0bb"} Oct 08 10:05:59 crc kubenswrapper[4744]: I1008 10:05:59.057722 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"1003e3fc-33a3-45d1-a1b7-9f383af6e336","Type":"ContainerStarted","Data":"dac09e347399a9cc6db3d54a5a5bb008036d5e10bbe421c86b387bf2466a9d9f"} Oct 08 10:05:59 crc kubenswrapper[4744]: I1008 10:05:59.068536 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" event={"ID":"65ab44b6-3478-42d3-b9cb-64f911f13342","Type":"ContainerDied","Data":"43b6ad8279ddf9acaed3feeeb223bde6e75cd46c627f21f2e95b393501fcabdf"} Oct 08 10:05:59 crc kubenswrapper[4744]: I1008 10:05:59.068607 4744 scope.go:117] "RemoveContainer" containerID="754ac2af91ca329df2d3d3340340cb5a2294499e668c5b00a0374f7e45b216dc" Oct 08 10:05:59 crc kubenswrapper[4744]: I1008 10:05:59.068793 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867c8fd5c5-wkswc" Oct 08 10:05:59 crc kubenswrapper[4744]: I1008 10:05:59.116547 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-867c8fd5c5-wkswc"] Oct 08 10:05:59 crc kubenswrapper[4744]: I1008 10:05:59.124877 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-867c8fd5c5-wkswc"] Oct 08 10:05:59 crc kubenswrapper[4744]: I1008 10:05:59.153141 4744 scope.go:117] "RemoveContainer" containerID="38f1147a1a639caf9a23723086d64ebcda76d5cd7c811418cdea6642b1368e6e" Oct 08 10:05:59 crc kubenswrapper[4744]: I1008 10:05:59.466465 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65ab44b6-3478-42d3-b9cb-64f911f13342" path="/var/lib/kubelet/pods/65ab44b6-3478-42d3-b9cb-64f911f13342/volumes" Oct 08 10:06:00 crc kubenswrapper[4744]: I1008 10:06:00.083850 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c","Type":"ContainerStarted","Data":"c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa"} Oct 08 10:06:00 crc kubenswrapper[4744]: I1008 10:06:00.091725 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"1003e3fc-33a3-45d1-a1b7-9f383af6e336","Type":"ContainerStarted","Data":"4effc120b1e16f9f8928f7e2072829a6a557b84c1f5ac0ea84281949349dd90e"} Oct 08 10:06:00 crc kubenswrapper[4744]: I1008 10:06:00.091835 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 08 10:06:00 crc kubenswrapper[4744]: I1008 10:06:00.111666 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=4.627609329 podStartE2EDuration="15.111646532s" podCreationTimestamp="2025-10-08 10:05:45 +0000 UTC" firstStartedPulling="2025-10-08 10:05:47.560587399 +0000 UTC m=+3242.808232638" lastFinishedPulling="2025-10-08 10:05:58.044624602 +0000 UTC m=+3253.292269841" observedRunningTime="2025-10-08 10:06:00.105474153 +0000 UTC m=+3255.353119392" watchObservedRunningTime="2025-10-08 10:06:00.111646532 +0000 UTC m=+3255.359291771" Oct 08 10:06:00 crc kubenswrapper[4744]: I1008 10:06:00.128233 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=8.128215147 podStartE2EDuration="8.128215147s" podCreationTimestamp="2025-10-08 10:05:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:06:00.121342798 +0000 UTC m=+3255.368988027" watchObservedRunningTime="2025-10-08 10:06:00.128215147 +0000 UTC m=+3255.375860386" Oct 08 10:06:00 crc kubenswrapper[4744]: I1008 10:06:00.555350 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 10:06:00 crc kubenswrapper[4744]: I1008 10:06:00.555944 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="sg-core" containerID="cri-o://291f257a2849b68458000e9bd8db98c8043e1f8bd5f6ebc02775d9dde2f6caf9" gracePeriod=30 Oct 08 10:06:00 crc kubenswrapper[4744]: I1008 10:06:00.555949 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="proxy-httpd" containerID="cri-o://ee689b75f227cdea446701892d4c60fef6c78f2ae31fae82c065c457d06e1169" gracePeriod=30 Oct 08 10:06:00 crc kubenswrapper[4744]: I1008 10:06:00.555979 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="ceilometer-notification-agent" containerID="cri-o://73753fe1ac100ccb50f054243e0ae9aca6a241a45c8569445513a0fbf6c18f32" gracePeriod=30 Oct 08 10:06:00 crc kubenswrapper[4744]: I1008 10:06:00.556309 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="ceilometer-central-agent" containerID="cri-o://dcfa0834c1f78532d0765c6898522eb0b5259139bd7ec1b122acddae4f284561" gracePeriod=30 Oct 08 10:06:01 crc kubenswrapper[4744]: I1008 10:06:01.106237 4744 generic.go:334] "Generic (PLEG): container finished" podID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerID="291f257a2849b68458000e9bd8db98c8043e1f8bd5f6ebc02775d9dde2f6caf9" exitCode=2 Oct 08 10:06:01 crc kubenswrapper[4744]: I1008 10:06:01.106277 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6a8bac2-8a56-45d4-a354-7814253f473d","Type":"ContainerDied","Data":"291f257a2849b68458000e9bd8db98c8043e1f8bd5f6ebc02775d9dde2f6caf9"} Oct 08 10:06:02 crc kubenswrapper[4744]: I1008 10:06:02.117941 4744 generic.go:334] "Generic (PLEG): container finished" podID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerID="ee689b75f227cdea446701892d4c60fef6c78f2ae31fae82c065c457d06e1169" exitCode=0 Oct 08 10:06:02 crc kubenswrapper[4744]: I1008 10:06:02.117971 4744 generic.go:334] "Generic (PLEG): container finished" podID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerID="dcfa0834c1f78532d0765c6898522eb0b5259139bd7ec1b122acddae4f284561" exitCode=0 Oct 08 10:06:02 crc kubenswrapper[4744]: I1008 10:06:02.117991 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6a8bac2-8a56-45d4-a354-7814253f473d","Type":"ContainerDied","Data":"ee689b75f227cdea446701892d4c60fef6c78f2ae31fae82c065c457d06e1169"} Oct 08 10:06:02 crc kubenswrapper[4744]: I1008 10:06:02.118020 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6a8bac2-8a56-45d4-a354-7814253f473d","Type":"ContainerDied","Data":"dcfa0834c1f78532d0765c6898522eb0b5259139bd7ec1b122acddae4f284561"} Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.116496 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.160955 4744 generic.go:334] "Generic (PLEG): container finished" podID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerID="73753fe1ac100ccb50f054243e0ae9aca6a241a45c8569445513a0fbf6c18f32" exitCode=0 Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.160998 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6a8bac2-8a56-45d4-a354-7814253f473d","Type":"ContainerDied","Data":"73753fe1ac100ccb50f054243e0ae9aca6a241a45c8569445513a0fbf6c18f32"} Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.310321 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.368142 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6a8bac2-8a56-45d4-a354-7814253f473d-log-httpd\") pod \"a6a8bac2-8a56-45d4-a354-7814253f473d\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.368203 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-config-data\") pod \"a6a8bac2-8a56-45d4-a354-7814253f473d\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.368239 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-ceilometer-tls-certs\") pod \"a6a8bac2-8a56-45d4-a354-7814253f473d\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.368338 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjxq2\" (UniqueName: \"kubernetes.io/projected/a6a8bac2-8a56-45d4-a354-7814253f473d-kube-api-access-gjxq2\") pod \"a6a8bac2-8a56-45d4-a354-7814253f473d\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.368377 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-scripts\") pod \"a6a8bac2-8a56-45d4-a354-7814253f473d\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.368423 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-sg-core-conf-yaml\") pod \"a6a8bac2-8a56-45d4-a354-7814253f473d\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.368439 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-combined-ca-bundle\") pod \"a6a8bac2-8a56-45d4-a354-7814253f473d\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.368471 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6a8bac2-8a56-45d4-a354-7814253f473d-run-httpd\") pod \"a6a8bac2-8a56-45d4-a354-7814253f473d\" (UID: \"a6a8bac2-8a56-45d4-a354-7814253f473d\") " Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.369502 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6a8bac2-8a56-45d4-a354-7814253f473d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a6a8bac2-8a56-45d4-a354-7814253f473d" (UID: "a6a8bac2-8a56-45d4-a354-7814253f473d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.369772 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6a8bac2-8a56-45d4-a354-7814253f473d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a6a8bac2-8a56-45d4-a354-7814253f473d" (UID: "a6a8bac2-8a56-45d4-a354-7814253f473d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.389682 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-scripts" (OuterVolumeSpecName: "scripts") pod "a6a8bac2-8a56-45d4-a354-7814253f473d" (UID: "a6a8bac2-8a56-45d4-a354-7814253f473d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.389851 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6a8bac2-8a56-45d4-a354-7814253f473d-kube-api-access-gjxq2" (OuterVolumeSpecName: "kube-api-access-gjxq2") pod "a6a8bac2-8a56-45d4-a354-7814253f473d" (UID: "a6a8bac2-8a56-45d4-a354-7814253f473d"). InnerVolumeSpecName "kube-api-access-gjxq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.426602 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a6a8bac2-8a56-45d4-a354-7814253f473d" (UID: "a6a8bac2-8a56-45d4-a354-7814253f473d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.470496 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6a8bac2-8a56-45d4-a354-7814253f473d-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.470533 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjxq2\" (UniqueName: \"kubernetes.io/projected/a6a8bac2-8a56-45d4-a354-7814253f473d-kube-api-access-gjxq2\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.470543 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.470552 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.470560 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a6a8bac2-8a56-45d4-a354-7814253f473d-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.477970 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a6a8bac2-8a56-45d4-a354-7814253f473d" (UID: "a6a8bac2-8a56-45d4-a354-7814253f473d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.527765 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6a8bac2-8a56-45d4-a354-7814253f473d" (UID: "a6a8bac2-8a56-45d4-a354-7814253f473d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.538928 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-config-data" (OuterVolumeSpecName: "config-data") pod "a6a8bac2-8a56-45d4-a354-7814253f473d" (UID: "a6a8bac2-8a56-45d4-a354-7814253f473d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.572747 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.572785 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:06 crc kubenswrapper[4744]: I1008 10:06:06.572798 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6a8bac2-8a56-45d4-a354-7814253f473d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.172653 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a6a8bac2-8a56-45d4-a354-7814253f473d","Type":"ContainerDied","Data":"9184172c78395e43513cdcde59ed391780fc131986ce75aaeb71e4c7cf33ad0b"} Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.172737 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.173083 4744 scope.go:117] "RemoveContainer" containerID="ee689b75f227cdea446701892d4c60fef6c78f2ae31fae82c065c457d06e1169" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.199793 4744 scope.go:117] "RemoveContainer" containerID="291f257a2849b68458000e9bd8db98c8043e1f8bd5f6ebc02775d9dde2f6caf9" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.203717 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.214924 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.234916 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 10:06:07 crc kubenswrapper[4744]: E1008 10:06:07.235271 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ab44b6-3478-42d3-b9cb-64f911f13342" containerName="dnsmasq-dns" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.235288 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ab44b6-3478-42d3-b9cb-64f911f13342" containerName="dnsmasq-dns" Oct 08 10:06:07 crc kubenswrapper[4744]: E1008 10:06:07.235297 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="sg-core" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.235303 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="sg-core" Oct 08 10:06:07 crc kubenswrapper[4744]: E1008 10:06:07.235319 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="ceilometer-central-agent" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.235325 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="ceilometer-central-agent" Oct 08 10:06:07 crc kubenswrapper[4744]: E1008 10:06:07.235337 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="proxy-httpd" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.235343 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="proxy-httpd" Oct 08 10:06:07 crc kubenswrapper[4744]: E1008 10:06:07.235355 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="ceilometer-notification-agent" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.235360 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="ceilometer-notification-agent" Oct 08 10:06:07 crc kubenswrapper[4744]: E1008 10:06:07.235389 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ab44b6-3478-42d3-b9cb-64f911f13342" containerName="init" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.235395 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ab44b6-3478-42d3-b9cb-64f911f13342" containerName="init" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.235561 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="proxy-httpd" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.235572 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ab44b6-3478-42d3-b9cb-64f911f13342" containerName="dnsmasq-dns" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.235582 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="sg-core" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.235592 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="ceilometer-central-agent" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.235638 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" containerName="ceilometer-notification-agent" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.237643 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.253319 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.305178 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.305418 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.305681 4744 scope.go:117] "RemoveContainer" containerID="73753fe1ac100ccb50f054243e0ae9aca6a241a45c8569445513a0fbf6c18f32" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.305925 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.313520 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.313643 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5224f\" (UniqueName: \"kubernetes.io/projected/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-kube-api-access-5224f\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.313666 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-run-httpd\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.313717 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-log-httpd\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.313802 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-config-data\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.313822 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.313988 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-scripts\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.314080 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.353708 4744 scope.go:117] "RemoveContainer" containerID="dcfa0834c1f78532d0765c6898522eb0b5259139bd7ec1b122acddae4f284561" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.416578 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.416645 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5224f\" (UniqueName: \"kubernetes.io/projected/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-kube-api-access-5224f\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.416663 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-run-httpd\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.416697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-log-httpd\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.416727 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-config-data\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.416745 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.416764 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-scripts\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.416788 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.417911 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-log-httpd\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.418544 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-run-httpd\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.422646 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.424089 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.431250 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-scripts\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.431782 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.434028 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5224f\" (UniqueName: \"kubernetes.io/projected/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-kube-api-access-5224f\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.434249 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2568ab3e-d5f6-4524-b1dc-20a6f389cd31-config-data\") pod \"ceilometer-0\" (UID: \"2568ab3e-d5f6-4524-b1dc-20a6f389cd31\") " pod="openstack/ceilometer-0" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.471980 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6a8bac2-8a56-45d4-a354-7814253f473d" path="/var/lib/kubelet/pods/a6a8bac2-8a56-45d4-a354-7814253f473d/volumes" Oct 08 10:06:07 crc kubenswrapper[4744]: I1008 10:06:07.626469 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 10:06:08 crc kubenswrapper[4744]: I1008 10:06:08.145990 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 10:06:08 crc kubenswrapper[4744]: I1008 10:06:08.181029 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2568ab3e-d5f6-4524-b1dc-20a6f389cd31","Type":"ContainerStarted","Data":"5355e1f0419b19defa90ac4ccfe16ec2a67bbc16757eb55f89e0f1ce238de572"} Oct 08 10:06:08 crc kubenswrapper[4744]: I1008 10:06:08.281200 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 08 10:06:08 crc kubenswrapper[4744]: I1008 10:06:08.319598 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 08 10:06:08 crc kubenswrapper[4744]: I1008 10:06:08.714237 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7f8f98d57b-66wrm" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Oct 08 10:06:09 crc kubenswrapper[4744]: I1008 10:06:09.191071 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="f80f1604-c09c-4ee6-bb49-b867311bf1e6" containerName="manila-scheduler" containerID="cri-o://e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3" gracePeriod=30 Oct 08 10:06:09 crc kubenswrapper[4744]: I1008 10:06:09.191452 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2568ab3e-d5f6-4524-b1dc-20a6f389cd31","Type":"ContainerStarted","Data":"08aeae738fc8c9baa5a15f6a060c91fb1651be902fa5b16b9939c44fe8dc4ce8"} Oct 08 10:06:09 crc kubenswrapper[4744]: I1008 10:06:09.191730 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="f80f1604-c09c-4ee6-bb49-b867311bf1e6" containerName="probe" containerID="cri-o://231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6" gracePeriod=30 Oct 08 10:06:10 crc kubenswrapper[4744]: I1008 10:06:10.201693 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2568ab3e-d5f6-4524-b1dc-20a6f389cd31","Type":"ContainerStarted","Data":"90b977f9f2f92497e316d567c9007172e69bc9175510b9114e86f8cbd951f788"} Oct 08 10:06:10 crc kubenswrapper[4744]: I1008 10:06:10.204877 4744 generic.go:334] "Generic (PLEG): container finished" podID="f80f1604-c09c-4ee6-bb49-b867311bf1e6" containerID="231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6" exitCode=0 Oct 08 10:06:10 crc kubenswrapper[4744]: I1008 10:06:10.204908 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f80f1604-c09c-4ee6-bb49-b867311bf1e6","Type":"ContainerDied","Data":"231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6"} Oct 08 10:06:10 crc kubenswrapper[4744]: I1008 10:06:10.453630 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:06:10 crc kubenswrapper[4744]: E1008 10:06:10.453951 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:06:10 crc kubenswrapper[4744]: I1008 10:06:10.867879 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 08 10:06:10 crc kubenswrapper[4744]: I1008 10:06:10.994682 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-combined-ca-bundle\") pod \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " Oct 08 10:06:10 crc kubenswrapper[4744]: I1008 10:06:10.994759 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-scripts\") pod \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " Oct 08 10:06:10 crc kubenswrapper[4744]: I1008 10:06:10.994809 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-config-data\") pod \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " Oct 08 10:06:10 crc kubenswrapper[4744]: I1008 10:06:10.994908 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f80f1604-c09c-4ee6-bb49-b867311bf1e6-etc-machine-id\") pod \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " Oct 08 10:06:10 crc kubenswrapper[4744]: I1008 10:06:10.995010 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grxq2\" (UniqueName: \"kubernetes.io/projected/f80f1604-c09c-4ee6-bb49-b867311bf1e6-kube-api-access-grxq2\") pod \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " Oct 08 10:06:10 crc kubenswrapper[4744]: I1008 10:06:10.995129 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-config-data-custom\") pod \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\" (UID: \"f80f1604-c09c-4ee6-bb49-b867311bf1e6\") " Oct 08 10:06:10 crc kubenswrapper[4744]: I1008 10:06:10.996780 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f80f1604-c09c-4ee6-bb49-b867311bf1e6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f80f1604-c09c-4ee6-bb49-b867311bf1e6" (UID: "f80f1604-c09c-4ee6-bb49-b867311bf1e6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.008955 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f80f1604-c09c-4ee6-bb49-b867311bf1e6-kube-api-access-grxq2" (OuterVolumeSpecName: "kube-api-access-grxq2") pod "f80f1604-c09c-4ee6-bb49-b867311bf1e6" (UID: "f80f1604-c09c-4ee6-bb49-b867311bf1e6"). InnerVolumeSpecName "kube-api-access-grxq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.012713 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f80f1604-c09c-4ee6-bb49-b867311bf1e6" (UID: "f80f1604-c09c-4ee6-bb49-b867311bf1e6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.016231 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-scripts" (OuterVolumeSpecName: "scripts") pod "f80f1604-c09c-4ee6-bb49-b867311bf1e6" (UID: "f80f1604-c09c-4ee6-bb49-b867311bf1e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.099262 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f80f1604-c09c-4ee6-bb49-b867311bf1e6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.099294 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grxq2\" (UniqueName: \"kubernetes.io/projected/f80f1604-c09c-4ee6-bb49-b867311bf1e6-kube-api-access-grxq2\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.099308 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.099318 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.104199 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f80f1604-c09c-4ee6-bb49-b867311bf1e6" (UID: "f80f1604-c09c-4ee6-bb49-b867311bf1e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.201557 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.223519 4744 generic.go:334] "Generic (PLEG): container finished" podID="f80f1604-c09c-4ee6-bb49-b867311bf1e6" containerID="e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3" exitCode=0 Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.223552 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f80f1604-c09c-4ee6-bb49-b867311bf1e6","Type":"ContainerDied","Data":"e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3"} Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.223599 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f80f1604-c09c-4ee6-bb49-b867311bf1e6","Type":"ContainerDied","Data":"f09d45847727726e2e322b87cd4717a99def8f9ecbd0c04acb83b161dbaf8760"} Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.223607 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.223617 4744 scope.go:117] "RemoveContainer" containerID="231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.223522 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-config-data" (OuterVolumeSpecName: "config-data") pod "f80f1604-c09c-4ee6-bb49-b867311bf1e6" (UID: "f80f1604-c09c-4ee6-bb49-b867311bf1e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.259997 4744 scope.go:117] "RemoveContainer" containerID="e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.284208 4744 scope.go:117] "RemoveContainer" containerID="231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6" Oct 08 10:06:11 crc kubenswrapper[4744]: E1008 10:06:11.285697 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6\": container with ID starting with 231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6 not found: ID does not exist" containerID="231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.285746 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6"} err="failed to get container status \"231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6\": rpc error: code = NotFound desc = could not find container \"231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6\": container with ID starting with 231db89cef2f3ad6c686fb31d1fc6894f4c70843ea3a786ded753894bc5844c6 not found: ID does not exist" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.285783 4744 scope.go:117] "RemoveContainer" containerID="e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3" Oct 08 10:06:11 crc kubenswrapper[4744]: E1008 10:06:11.286186 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3\": container with ID starting with e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3 not found: ID does not exist" containerID="e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.286232 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3"} err="failed to get container status \"e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3\": rpc error: code = NotFound desc = could not find container \"e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3\": container with ID starting with e5f6bb76db6d1ad864befb6bd36b114354569e9841d734ea84620b54064c35b3 not found: ID does not exist" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.304069 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f80f1604-c09c-4ee6-bb49-b867311bf1e6-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.563602 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.576728 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.587924 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 08 10:06:11 crc kubenswrapper[4744]: E1008 10:06:11.588560 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f80f1604-c09c-4ee6-bb49-b867311bf1e6" containerName="probe" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.588657 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f80f1604-c09c-4ee6-bb49-b867311bf1e6" containerName="probe" Oct 08 10:06:11 crc kubenswrapper[4744]: E1008 10:06:11.588745 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f80f1604-c09c-4ee6-bb49-b867311bf1e6" containerName="manila-scheduler" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.588821 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f80f1604-c09c-4ee6-bb49-b867311bf1e6" containerName="manila-scheduler" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.589060 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f80f1604-c09c-4ee6-bb49-b867311bf1e6" containerName="manila-scheduler" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.589140 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f80f1604-c09c-4ee6-bb49-b867311bf1e6" containerName="probe" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.590122 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.597382 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.599671 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.721960 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffw2l\" (UniqueName: \"kubernetes.io/projected/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-kube-api-access-ffw2l\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.722012 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-config-data\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.722050 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.722266 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.722621 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.722787 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-scripts\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.824281 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.824407 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.824451 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-scripts\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.824537 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffw2l\" (UniqueName: \"kubernetes.io/projected/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-kube-api-access-ffw2l\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.824568 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-config-data\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.824616 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.824609 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.829128 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-scripts\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.829857 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.831122 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.837596 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-config-data\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.856835 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffw2l\" (UniqueName: \"kubernetes.io/projected/3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e-kube-api-access-ffw2l\") pod \"manila-scheduler-0\" (UID: \"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e\") " pod="openstack/manila-scheduler-0" Oct 08 10:06:11 crc kubenswrapper[4744]: I1008 10:06:11.920993 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 08 10:06:12 crc kubenswrapper[4744]: I1008 10:06:12.238511 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2568ab3e-d5f6-4524-b1dc-20a6f389cd31","Type":"ContainerStarted","Data":"eef0c3441ed0cd979c1f0c12fe47ac43b1ec54f8b7c1866ed8f1c5e140b6fcd6"} Oct 08 10:06:12 crc kubenswrapper[4744]: I1008 10:06:12.405268 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 08 10:06:13 crc kubenswrapper[4744]: I1008 10:06:13.270259 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e","Type":"ContainerStarted","Data":"c85760a12f8830bee345eee4cabcaec604e9d5c567002ad2466fb63037d8ddfe"} Oct 08 10:06:13 crc kubenswrapper[4744]: I1008 10:06:13.270695 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e","Type":"ContainerStarted","Data":"a7c923fec86ec7ec307a6299fe72066f4ab00f18f18082529199b3605529b443"} Oct 08 10:06:13 crc kubenswrapper[4744]: I1008 10:06:13.282166 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2568ab3e-d5f6-4524-b1dc-20a6f389cd31","Type":"ContainerStarted","Data":"0804198824f0cca785108a2465a3dc4c10bad736c1af97aa7a07fbafc55a4efb"} Oct 08 10:06:13 crc kubenswrapper[4744]: I1008 10:06:13.282363 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 10:06:13 crc kubenswrapper[4744]: I1008 10:06:13.314091 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.990273846 podStartE2EDuration="6.314063748s" podCreationTimestamp="2025-10-08 10:06:07 +0000 UTC" firstStartedPulling="2025-10-08 10:06:08.147930557 +0000 UTC m=+3263.395575796" lastFinishedPulling="2025-10-08 10:06:12.471720469 +0000 UTC m=+3267.719365698" observedRunningTime="2025-10-08 10:06:13.309110222 +0000 UTC m=+3268.556755461" watchObservedRunningTime="2025-10-08 10:06:13.314063748 +0000 UTC m=+3268.561708977" Oct 08 10:06:13 crc kubenswrapper[4744]: I1008 10:06:13.467535 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f80f1604-c09c-4ee6-bb49-b867311bf1e6" path="/var/lib/kubelet/pods/f80f1604-c09c-4ee6-bb49-b867311bf1e6/volumes" Oct 08 10:06:14 crc kubenswrapper[4744]: I1008 10:06:14.292586 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e","Type":"ContainerStarted","Data":"53f6d8270db4acdde6f895300dd1ddee4b01cd0f7091abacf00e8f22f9a13386"} Oct 08 10:06:14 crc kubenswrapper[4744]: I1008 10:06:14.321026 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.321009679 podStartE2EDuration="3.321009679s" podCreationTimestamp="2025-10-08 10:06:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:06:14.311080277 +0000 UTC m=+3269.558725516" watchObservedRunningTime="2025-10-08 10:06:14.321009679 +0000 UTC m=+3269.568654918" Oct 08 10:06:14 crc kubenswrapper[4744]: I1008 10:06:14.906061 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 08 10:06:17 crc kubenswrapper[4744]: I1008 10:06:17.643909 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 08 10:06:17 crc kubenswrapper[4744]: I1008 10:06:17.719089 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 08 10:06:18 crc kubenswrapper[4744]: I1008 10:06:18.329956 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" containerName="manila-share" containerID="cri-o://f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd" gracePeriod=30 Oct 08 10:06:18 crc kubenswrapper[4744]: I1008 10:06:18.329987 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" containerName="probe" containerID="cri-o://c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa" gracePeriod=30 Oct 08 10:06:18 crc kubenswrapper[4744]: I1008 10:06:18.714109 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7f8f98d57b-66wrm" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.263757 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.343346 4744 generic.go:334] "Generic (PLEG): container finished" podID="2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" containerID="c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa" exitCode=0 Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.343388 4744 generic.go:334] "Generic (PLEG): container finished" podID="2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" containerID="f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd" exitCode=1 Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.343405 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c","Type":"ContainerDied","Data":"c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa"} Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.343429 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c","Type":"ContainerDied","Data":"f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd"} Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.343440 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c","Type":"ContainerDied","Data":"1dac840eabf2adfa1b9b91877b83b6eb2446c4b25bc8627eed40ddf4f49fd4e4"} Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.343454 4744 scope.go:117] "RemoveContainer" containerID="c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.343558 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.371864 4744 scope.go:117] "RemoveContainer" containerID="f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.384024 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-var-lib-manila\") pod \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.384407 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-config-data-custom\") pod \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.384433 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-ceph\") pod \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.384484 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88tlb\" (UniqueName: \"kubernetes.io/projected/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-kube-api-access-88tlb\") pod \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.384506 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-config-data\") pod \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.384559 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-etc-machine-id\") pod \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.384692 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-combined-ca-bundle\") pod \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.384127 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" (UID: "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.384760 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-scripts\") pod \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\" (UID: \"2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c\") " Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.384805 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" (UID: "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.385385 4744 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-var-lib-manila\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.385401 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.393336 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" (UID: "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.393773 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-ceph" (OuterVolumeSpecName: "ceph") pod "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" (UID: "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.396569 4744 scope.go:117] "RemoveContainer" containerID="c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa" Oct 08 10:06:19 crc kubenswrapper[4744]: E1008 10:06:19.399067 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa\": container with ID starting with c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa not found: ID does not exist" containerID="c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.399217 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa"} err="failed to get container status \"c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa\": rpc error: code = NotFound desc = could not find container \"c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa\": container with ID starting with c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa not found: ID does not exist" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.399331 4744 scope.go:117] "RemoveContainer" containerID="f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.399619 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-kube-api-access-88tlb" (OuterVolumeSpecName: "kube-api-access-88tlb") pod "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" (UID: "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c"). InnerVolumeSpecName "kube-api-access-88tlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:06:19 crc kubenswrapper[4744]: E1008 10:06:19.403542 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd\": container with ID starting with f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd not found: ID does not exist" containerID="f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.403604 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd"} err="failed to get container status \"f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd\": rpc error: code = NotFound desc = could not find container \"f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd\": container with ID starting with f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd not found: ID does not exist" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.403635 4744 scope.go:117] "RemoveContainer" containerID="c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.408494 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa"} err="failed to get container status \"c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa\": rpc error: code = NotFound desc = could not find container \"c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa\": container with ID starting with c3e5ca637261327c2c347f03939b13af84e7318d41f881d64f1afb3b85cb8bfa not found: ID does not exist" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.408533 4744 scope.go:117] "RemoveContainer" containerID="f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.409615 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd"} err="failed to get container status \"f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd\": rpc error: code = NotFound desc = could not find container \"f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd\": container with ID starting with f242d34ab632bedbbcb9eaa7c6155e3bf370481c44ea11a46a8f56fd3d78d2bd not found: ID does not exist" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.412738 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-scripts" (OuterVolumeSpecName: "scripts") pod "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" (UID: "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.476543 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" (UID: "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.487461 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.487496 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.487509 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-ceph\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.487518 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88tlb\" (UniqueName: \"kubernetes.io/projected/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-kube-api-access-88tlb\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.487527 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.525674 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-config-data" (OuterVolumeSpecName: "config-data") pod "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" (UID: "2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.589722 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.680517 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.690803 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.721957 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 08 10:06:19 crc kubenswrapper[4744]: E1008 10:06:19.722612 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" containerName="probe" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.722633 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" containerName="probe" Oct 08 10:06:19 crc kubenswrapper[4744]: E1008 10:06:19.722660 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" containerName="manila-share" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.722668 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" containerName="manila-share" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.722870 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" containerName="manila-share" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.722888 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" containerName="probe" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.723897 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.727955 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.755985 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.793981 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0e308a1c-d62d-4782-adba-7517aab54f94-ceph\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.794039 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e308a1c-d62d-4782-adba-7517aab54f94-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.794089 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/0e308a1c-d62d-4782-adba-7517aab54f94-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.794152 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e308a1c-d62d-4782-adba-7517aab54f94-scripts\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.794180 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e308a1c-d62d-4782-adba-7517aab54f94-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.794223 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e308a1c-d62d-4782-adba-7517aab54f94-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.794255 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e308a1c-d62d-4782-adba-7517aab54f94-config-data\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.794310 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm4k5\" (UniqueName: \"kubernetes.io/projected/0e308a1c-d62d-4782-adba-7517aab54f94-kube-api-access-cm4k5\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.896834 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e308a1c-d62d-4782-adba-7517aab54f94-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.896919 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e308a1c-d62d-4782-adba-7517aab54f94-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.896965 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e308a1c-d62d-4782-adba-7517aab54f94-config-data\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.897071 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm4k5\" (UniqueName: \"kubernetes.io/projected/0e308a1c-d62d-4782-adba-7517aab54f94-kube-api-access-cm4k5\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.897131 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0e308a1c-d62d-4782-adba-7517aab54f94-ceph\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.897159 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e308a1c-d62d-4782-adba-7517aab54f94-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.897199 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/0e308a1c-d62d-4782-adba-7517aab54f94-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.897254 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e308a1c-d62d-4782-adba-7517aab54f94-scripts\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.897695 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e308a1c-d62d-4782-adba-7517aab54f94-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.897781 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/0e308a1c-d62d-4782-adba-7517aab54f94-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.901028 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e308a1c-d62d-4782-adba-7517aab54f94-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.901161 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e308a1c-d62d-4782-adba-7517aab54f94-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.901596 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0e308a1c-d62d-4782-adba-7517aab54f94-ceph\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.902561 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e308a1c-d62d-4782-adba-7517aab54f94-scripts\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.906634 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e308a1c-d62d-4782-adba-7517aab54f94-config-data\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:19 crc kubenswrapper[4744]: I1008 10:06:19.917025 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm4k5\" (UniqueName: \"kubernetes.io/projected/0e308a1c-d62d-4782-adba-7517aab54f94-kube-api-access-cm4k5\") pod \"manila-share-share1-0\" (UID: \"0e308a1c-d62d-4782-adba-7517aab54f94\") " pod="openstack/manila-share-share1-0" Oct 08 10:06:20 crc kubenswrapper[4744]: I1008 10:06:20.103834 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 08 10:06:20 crc kubenswrapper[4744]: I1008 10:06:20.698881 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 08 10:06:21 crc kubenswrapper[4744]: I1008 10:06:21.366575 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"0e308a1c-d62d-4782-adba-7517aab54f94","Type":"ContainerStarted","Data":"f100bfee3a0d0661e6e6fa20387508a4f6036160cdd6a7a2de512c938ed4d37b"} Oct 08 10:06:21 crc kubenswrapper[4744]: I1008 10:06:21.366892 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"0e308a1c-d62d-4782-adba-7517aab54f94","Type":"ContainerStarted","Data":"cf0eb404e069602bc9025678fbeab452980b0ab82f44b0f978e9a6fb58f15947"} Oct 08 10:06:21 crc kubenswrapper[4744]: I1008 10:06:21.465343 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c" path="/var/lib/kubelet/pods/2d89cc63-9062-4f6f-aa7f-f8a0ebfb8f6c/volumes" Oct 08 10:06:21 crc kubenswrapper[4744]: I1008 10:06:21.921884 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 08 10:06:22 crc kubenswrapper[4744]: I1008 10:06:22.379446 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"0e308a1c-d62d-4782-adba-7517aab54f94","Type":"ContainerStarted","Data":"1700140fd5d5e24f02976553d7b7e3cf2e3b27704ab0b79e1a1e7e8aa126b606"} Oct 08 10:06:22 crc kubenswrapper[4744]: I1008 10:06:22.413171 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.413147426 podStartE2EDuration="3.413147426s" podCreationTimestamp="2025-10-08 10:06:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:06:22.406063811 +0000 UTC m=+3277.653709050" watchObservedRunningTime="2025-10-08 10:06:22.413147426 +0000 UTC m=+3277.660792655" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.421615 4744 generic.go:334] "Generic (PLEG): container finished" podID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerID="621e50ea58c51b02bda6c75857ac031b6d770e3776c4efe36a0937f13140b393" exitCode=137 Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.422084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f8f98d57b-66wrm" event={"ID":"49150e8b-4e4b-467f-ae7b-a6d14b63554d","Type":"ContainerDied","Data":"621e50ea58c51b02bda6c75857ac031b6d770e3776c4efe36a0937f13140b393"} Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.460504 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:06:25 crc kubenswrapper[4744]: E1008 10:06:25.460789 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.592617 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.655355 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbgfv\" (UniqueName: \"kubernetes.io/projected/49150e8b-4e4b-467f-ae7b-a6d14b63554d-kube-api-access-hbgfv\") pod \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.655464 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-combined-ca-bundle\") pod \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.655506 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-horizon-secret-key\") pod \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.655530 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-horizon-tls-certs\") pod \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.655624 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49150e8b-4e4b-467f-ae7b-a6d14b63554d-config-data\") pod \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.655774 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49150e8b-4e4b-467f-ae7b-a6d14b63554d-scripts\") pod \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.655860 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49150e8b-4e4b-467f-ae7b-a6d14b63554d-logs\") pod \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\" (UID: \"49150e8b-4e4b-467f-ae7b-a6d14b63554d\") " Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.656739 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49150e8b-4e4b-467f-ae7b-a6d14b63554d-logs" (OuterVolumeSpecName: "logs") pod "49150e8b-4e4b-467f-ae7b-a6d14b63554d" (UID: "49150e8b-4e4b-467f-ae7b-a6d14b63554d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.668718 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "49150e8b-4e4b-467f-ae7b-a6d14b63554d" (UID: "49150e8b-4e4b-467f-ae7b-a6d14b63554d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.668720 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49150e8b-4e4b-467f-ae7b-a6d14b63554d-kube-api-access-hbgfv" (OuterVolumeSpecName: "kube-api-access-hbgfv") pod "49150e8b-4e4b-467f-ae7b-a6d14b63554d" (UID: "49150e8b-4e4b-467f-ae7b-a6d14b63554d"). InnerVolumeSpecName "kube-api-access-hbgfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.687839 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49150e8b-4e4b-467f-ae7b-a6d14b63554d-scripts" (OuterVolumeSpecName: "scripts") pod "49150e8b-4e4b-467f-ae7b-a6d14b63554d" (UID: "49150e8b-4e4b-467f-ae7b-a6d14b63554d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.700544 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49150e8b-4e4b-467f-ae7b-a6d14b63554d" (UID: "49150e8b-4e4b-467f-ae7b-a6d14b63554d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.716175 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49150e8b-4e4b-467f-ae7b-a6d14b63554d-config-data" (OuterVolumeSpecName: "config-data") pod "49150e8b-4e4b-467f-ae7b-a6d14b63554d" (UID: "49150e8b-4e4b-467f-ae7b-a6d14b63554d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.720205 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "49150e8b-4e4b-467f-ae7b-a6d14b63554d" (UID: "49150e8b-4e4b-467f-ae7b-a6d14b63554d"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.758317 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49150e8b-4e4b-467f-ae7b-a6d14b63554d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.758360 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49150e8b-4e4b-467f-ae7b-a6d14b63554d-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.758388 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49150e8b-4e4b-467f-ae7b-a6d14b63554d-logs\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.758398 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbgfv\" (UniqueName: \"kubernetes.io/projected/49150e8b-4e4b-467f-ae7b-a6d14b63554d-kube-api-access-hbgfv\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.758410 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.758420 4744 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:25 crc kubenswrapper[4744]: I1008 10:06:25.758428 4744 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/49150e8b-4e4b-467f-ae7b-a6d14b63554d-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 10:06:26 crc kubenswrapper[4744]: I1008 10:06:26.446254 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f8f98d57b-66wrm" event={"ID":"49150e8b-4e4b-467f-ae7b-a6d14b63554d","Type":"ContainerDied","Data":"0212bbd8f0391c84b8a5e18e4f583c8b159752be2f8fc1075d40ddb0783a5029"} Oct 08 10:06:26 crc kubenswrapper[4744]: I1008 10:06:26.446309 4744 scope.go:117] "RemoveContainer" containerID="2139c1999ee135e2b62d6e294e9ec99ea5173c7f974b9cfaa2786d28f6dac0bb" Oct 08 10:06:26 crc kubenswrapper[4744]: I1008 10:06:26.446397 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f8f98d57b-66wrm" Oct 08 10:06:26 crc kubenswrapper[4744]: I1008 10:06:26.495171 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f8f98d57b-66wrm"] Oct 08 10:06:26 crc kubenswrapper[4744]: I1008 10:06:26.507205 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f8f98d57b-66wrm"] Oct 08 10:06:26 crc kubenswrapper[4744]: I1008 10:06:26.618092 4744 scope.go:117] "RemoveContainer" containerID="621e50ea58c51b02bda6c75857ac031b6d770e3776c4efe36a0937f13140b393" Oct 08 10:06:27 crc kubenswrapper[4744]: I1008 10:06:27.471856 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" path="/var/lib/kubelet/pods/49150e8b-4e4b-467f-ae7b-a6d14b63554d/volumes" Oct 08 10:06:30 crc kubenswrapper[4744]: I1008 10:06:30.104856 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 08 10:06:33 crc kubenswrapper[4744]: I1008 10:06:33.580348 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 08 10:06:37 crc kubenswrapper[4744]: I1008 10:06:37.637685 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 10:06:39 crc kubenswrapper[4744]: I1008 10:06:39.453846 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:06:39 crc kubenswrapper[4744]: E1008 10:06:39.454459 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:06:41 crc kubenswrapper[4744]: I1008 10:06:41.787577 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 08 10:06:51 crc kubenswrapper[4744]: I1008 10:06:51.453190 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:06:51 crc kubenswrapper[4744]: I1008 10:06:51.736023 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"24813bdc7353e22419b64c64e6f4e18c33e77fd646809a089b4e5a14586c42b3"} Oct 08 10:07:33 crc kubenswrapper[4744]: I1008 10:07:33.750512 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-scl94"] Oct 08 10:07:33 crc kubenswrapper[4744]: E1008 10:07:33.751498 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon" Oct 08 10:07:33 crc kubenswrapper[4744]: I1008 10:07:33.751511 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon" Oct 08 10:07:33 crc kubenswrapper[4744]: E1008 10:07:33.751532 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon-log" Oct 08 10:07:33 crc kubenswrapper[4744]: I1008 10:07:33.751539 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon-log" Oct 08 10:07:33 crc kubenswrapper[4744]: I1008 10:07:33.751728 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon-log" Oct 08 10:07:33 crc kubenswrapper[4744]: I1008 10:07:33.751744 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="49150e8b-4e4b-467f-ae7b-a6d14b63554d" containerName="horizon" Oct 08 10:07:33 crc kubenswrapper[4744]: I1008 10:07:33.753158 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:33 crc kubenswrapper[4744]: I1008 10:07:33.766175 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-scl94"] Oct 08 10:07:33 crc kubenswrapper[4744]: I1008 10:07:33.910547 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee7f1dc6-9467-488a-9da6-947f88cb02ac-catalog-content\") pod \"redhat-operators-scl94\" (UID: \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\") " pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:33 crc kubenswrapper[4744]: I1008 10:07:33.910696 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee7f1dc6-9467-488a-9da6-947f88cb02ac-utilities\") pod \"redhat-operators-scl94\" (UID: \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\") " pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:33 crc kubenswrapper[4744]: I1008 10:07:33.910946 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br9wz\" (UniqueName: \"kubernetes.io/projected/ee7f1dc6-9467-488a-9da6-947f88cb02ac-kube-api-access-br9wz\") pod \"redhat-operators-scl94\" (UID: \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\") " pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:34 crc kubenswrapper[4744]: I1008 10:07:34.012849 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee7f1dc6-9467-488a-9da6-947f88cb02ac-catalog-content\") pod \"redhat-operators-scl94\" (UID: \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\") " pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:34 crc kubenswrapper[4744]: I1008 10:07:34.012934 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee7f1dc6-9467-488a-9da6-947f88cb02ac-utilities\") pod \"redhat-operators-scl94\" (UID: \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\") " pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:34 crc kubenswrapper[4744]: I1008 10:07:34.013003 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br9wz\" (UniqueName: \"kubernetes.io/projected/ee7f1dc6-9467-488a-9da6-947f88cb02ac-kube-api-access-br9wz\") pod \"redhat-operators-scl94\" (UID: \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\") " pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:34 crc kubenswrapper[4744]: I1008 10:07:34.013825 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee7f1dc6-9467-488a-9da6-947f88cb02ac-catalog-content\") pod \"redhat-operators-scl94\" (UID: \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\") " pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:34 crc kubenswrapper[4744]: I1008 10:07:34.014042 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee7f1dc6-9467-488a-9da6-947f88cb02ac-utilities\") pod \"redhat-operators-scl94\" (UID: \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\") " pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:34 crc kubenswrapper[4744]: I1008 10:07:34.045324 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br9wz\" (UniqueName: \"kubernetes.io/projected/ee7f1dc6-9467-488a-9da6-947f88cb02ac-kube-api-access-br9wz\") pod \"redhat-operators-scl94\" (UID: \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\") " pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:34 crc kubenswrapper[4744]: I1008 10:07:34.074105 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:34 crc kubenswrapper[4744]: I1008 10:07:34.515819 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-scl94"] Oct 08 10:07:35 crc kubenswrapper[4744]: I1008 10:07:35.205167 4744 generic.go:334] "Generic (PLEG): container finished" podID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerID="870a934102c8ce02794d9d65f581e1d5f1df0176940d012e2427a4f3324a4212" exitCode=0 Oct 08 10:07:35 crc kubenswrapper[4744]: I1008 10:07:35.205631 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scl94" event={"ID":"ee7f1dc6-9467-488a-9da6-947f88cb02ac","Type":"ContainerDied","Data":"870a934102c8ce02794d9d65f581e1d5f1df0176940d012e2427a4f3324a4212"} Oct 08 10:07:35 crc kubenswrapper[4744]: I1008 10:07:35.205673 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scl94" event={"ID":"ee7f1dc6-9467-488a-9da6-947f88cb02ac","Type":"ContainerStarted","Data":"8f7b0b6a52e5f618fd51df66374c2243128917d1b680af9d73cac753ac6e8bf5"} Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.232365 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scl94" event={"ID":"ee7f1dc6-9467-488a-9da6-947f88cb02ac","Type":"ContainerStarted","Data":"b71b493ec4b83358c9f7e14a7a85dc52efc7eff769558b4b3e396b7ffe670bd8"} Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.448658 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.450741 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.479234 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.479339 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-g5r6h" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.481907 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.481960 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.488092 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.614142 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.614494 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.614602 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.614727 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.614811 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.614847 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h776v\" (UniqueName: \"kubernetes.io/projected/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-kube-api-access-h776v\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.614915 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-config-data\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.615192 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.615691 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.717631 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h776v\" (UniqueName: \"kubernetes.io/projected/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-kube-api-access-h776v\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.717689 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-config-data\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.717742 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.717825 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.717871 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.717890 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.717920 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.717958 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.718000 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.718595 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.719150 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.719406 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-config-data\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.719719 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.724148 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.725954 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.733146 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.744116 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h776v\" (UniqueName: \"kubernetes.io/projected/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-kube-api-access-h776v\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.751682 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.788226 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " pod="openstack/tempest-tests-tempest" Oct 08 10:07:37 crc kubenswrapper[4744]: I1008 10:07:37.816650 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 10:07:38 crc kubenswrapper[4744]: I1008 10:07:38.317523 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 08 10:07:39 crc kubenswrapper[4744]: I1008 10:07:39.248980 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be","Type":"ContainerStarted","Data":"fd623e980c7ee1817d3a8eb3ad7c3aa7e53ed23408f0966446fbe864a4afea02"} Oct 08 10:07:42 crc kubenswrapper[4744]: I1008 10:07:42.293594 4744 generic.go:334] "Generic (PLEG): container finished" podID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerID="b71b493ec4b83358c9f7e14a7a85dc52efc7eff769558b4b3e396b7ffe670bd8" exitCode=0 Oct 08 10:07:42 crc kubenswrapper[4744]: I1008 10:07:42.293677 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scl94" event={"ID":"ee7f1dc6-9467-488a-9da6-947f88cb02ac","Type":"ContainerDied","Data":"b71b493ec4b83358c9f7e14a7a85dc52efc7eff769558b4b3e396b7ffe670bd8"} Oct 08 10:07:47 crc kubenswrapper[4744]: I1008 10:07:47.362929 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scl94" event={"ID":"ee7f1dc6-9467-488a-9da6-947f88cb02ac","Type":"ContainerStarted","Data":"9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39"} Oct 08 10:07:47 crc kubenswrapper[4744]: I1008 10:07:47.383919 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-scl94" podStartSLOduration=2.9568758280000003 podStartE2EDuration="14.38389938s" podCreationTimestamp="2025-10-08 10:07:33 +0000 UTC" firstStartedPulling="2025-10-08 10:07:35.210940084 +0000 UTC m=+3350.458585323" lastFinishedPulling="2025-10-08 10:07:46.637963636 +0000 UTC m=+3361.885608875" observedRunningTime="2025-10-08 10:07:47.382529923 +0000 UTC m=+3362.630175212" watchObservedRunningTime="2025-10-08 10:07:47.38389938 +0000 UTC m=+3362.631544609" Oct 08 10:07:48 crc kubenswrapper[4744]: I1008 10:07:48.854709 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rd7n4"] Oct 08 10:07:48 crc kubenswrapper[4744]: I1008 10:07:48.856686 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:48 crc kubenswrapper[4744]: I1008 10:07:48.870880 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rd7n4"] Oct 08 10:07:48 crc kubenswrapper[4744]: I1008 10:07:48.995702 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b76beff0-6871-4de0-a50c-82327bae3b2f-catalog-content\") pod \"redhat-marketplace-rd7n4\" (UID: \"b76beff0-6871-4de0-a50c-82327bae3b2f\") " pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:48 crc kubenswrapper[4744]: I1008 10:07:48.996036 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b76beff0-6871-4de0-a50c-82327bae3b2f-utilities\") pod \"redhat-marketplace-rd7n4\" (UID: \"b76beff0-6871-4de0-a50c-82327bae3b2f\") " pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:48 crc kubenswrapper[4744]: I1008 10:07:48.996175 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jchrh\" (UniqueName: \"kubernetes.io/projected/b76beff0-6871-4de0-a50c-82327bae3b2f-kube-api-access-jchrh\") pod \"redhat-marketplace-rd7n4\" (UID: \"b76beff0-6871-4de0-a50c-82327bae3b2f\") " pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:49 crc kubenswrapper[4744]: I1008 10:07:49.098443 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b76beff0-6871-4de0-a50c-82327bae3b2f-catalog-content\") pod \"redhat-marketplace-rd7n4\" (UID: \"b76beff0-6871-4de0-a50c-82327bae3b2f\") " pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:49 crc kubenswrapper[4744]: I1008 10:07:49.098499 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b76beff0-6871-4de0-a50c-82327bae3b2f-utilities\") pod \"redhat-marketplace-rd7n4\" (UID: \"b76beff0-6871-4de0-a50c-82327bae3b2f\") " pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:49 crc kubenswrapper[4744]: I1008 10:07:49.098569 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jchrh\" (UniqueName: \"kubernetes.io/projected/b76beff0-6871-4de0-a50c-82327bae3b2f-kube-api-access-jchrh\") pod \"redhat-marketplace-rd7n4\" (UID: \"b76beff0-6871-4de0-a50c-82327bae3b2f\") " pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:49 crc kubenswrapper[4744]: I1008 10:07:49.099431 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b76beff0-6871-4de0-a50c-82327bae3b2f-catalog-content\") pod \"redhat-marketplace-rd7n4\" (UID: \"b76beff0-6871-4de0-a50c-82327bae3b2f\") " pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:49 crc kubenswrapper[4744]: I1008 10:07:49.099634 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b76beff0-6871-4de0-a50c-82327bae3b2f-utilities\") pod \"redhat-marketplace-rd7n4\" (UID: \"b76beff0-6871-4de0-a50c-82327bae3b2f\") " pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:49 crc kubenswrapper[4744]: I1008 10:07:49.120754 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jchrh\" (UniqueName: \"kubernetes.io/projected/b76beff0-6871-4de0-a50c-82327bae3b2f-kube-api-access-jchrh\") pod \"redhat-marketplace-rd7n4\" (UID: \"b76beff0-6871-4de0-a50c-82327bae3b2f\") " pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:49 crc kubenswrapper[4744]: I1008 10:07:49.182615 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:49 crc kubenswrapper[4744]: I1008 10:07:49.744281 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rd7n4"] Oct 08 10:07:50 crc kubenswrapper[4744]: I1008 10:07:50.409418 4744 generic.go:334] "Generic (PLEG): container finished" podID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerID="d2f8a6f37511f36d941635a9fee729c7cb8a1affbecaf7fa32b3dc292cc3c4f1" exitCode=0 Oct 08 10:07:50 crc kubenswrapper[4744]: I1008 10:07:50.409662 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd7n4" event={"ID":"b76beff0-6871-4de0-a50c-82327bae3b2f","Type":"ContainerDied","Data":"d2f8a6f37511f36d941635a9fee729c7cb8a1affbecaf7fa32b3dc292cc3c4f1"} Oct 08 10:07:50 crc kubenswrapper[4744]: I1008 10:07:50.409774 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd7n4" event={"ID":"b76beff0-6871-4de0-a50c-82327bae3b2f","Type":"ContainerStarted","Data":"2857add3449a6fa083596bf36239000821981c4b50433f6e6c69710c0ded6e20"} Oct 08 10:07:51 crc kubenswrapper[4744]: I1008 10:07:51.426063 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd7n4" event={"ID":"b76beff0-6871-4de0-a50c-82327bae3b2f","Type":"ContainerStarted","Data":"cfab3e2f252a3a52d4b5067160ecb642e8301245ac6333211c98e130d7868697"} Oct 08 10:07:53 crc kubenswrapper[4744]: I1008 10:07:53.451544 4744 generic.go:334] "Generic (PLEG): container finished" podID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerID="cfab3e2f252a3a52d4b5067160ecb642e8301245ac6333211c98e130d7868697" exitCode=0 Oct 08 10:07:53 crc kubenswrapper[4744]: I1008 10:07:53.451940 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd7n4" event={"ID":"b76beff0-6871-4de0-a50c-82327bae3b2f","Type":"ContainerDied","Data":"cfab3e2f252a3a52d4b5067160ecb642e8301245ac6333211c98e130d7868697"} Oct 08 10:07:54 crc kubenswrapper[4744]: I1008 10:07:54.074485 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:54 crc kubenswrapper[4744]: I1008 10:07:54.074807 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:07:54 crc kubenswrapper[4744]: I1008 10:07:54.462900 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd7n4" event={"ID":"b76beff0-6871-4de0-a50c-82327bae3b2f","Type":"ContainerStarted","Data":"2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944"} Oct 08 10:07:54 crc kubenswrapper[4744]: I1008 10:07:54.481297 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rd7n4" podStartSLOduration=3.020405207 podStartE2EDuration="6.481277408s" podCreationTimestamp="2025-10-08 10:07:48 +0000 UTC" firstStartedPulling="2025-10-08 10:07:50.412614101 +0000 UTC m=+3365.660259340" lastFinishedPulling="2025-10-08 10:07:53.873486302 +0000 UTC m=+3369.121131541" observedRunningTime="2025-10-08 10:07:54.479257653 +0000 UTC m=+3369.726902902" watchObservedRunningTime="2025-10-08 10:07:54.481277408 +0000 UTC m=+3369.728922647" Oct 08 10:07:55 crc kubenswrapper[4744]: I1008 10:07:55.128136 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-scl94" podUID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerName="registry-server" probeResult="failure" output=< Oct 08 10:07:55 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 10:07:55 crc kubenswrapper[4744]: > Oct 08 10:07:59 crc kubenswrapper[4744]: I1008 10:07:59.182984 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:59 crc kubenswrapper[4744]: I1008 10:07:59.183508 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:59 crc kubenswrapper[4744]: I1008 10:07:59.234961 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:59 crc kubenswrapper[4744]: I1008 10:07:59.620899 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:07:59 crc kubenswrapper[4744]: I1008 10:07:59.667490 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rd7n4"] Oct 08 10:08:01 crc kubenswrapper[4744]: I1008 10:08:01.542261 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rd7n4" podUID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerName="registry-server" containerID="cri-o://2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944" gracePeriod=2 Oct 08 10:08:02 crc kubenswrapper[4744]: I1008 10:08:02.562805 4744 generic.go:334] "Generic (PLEG): container finished" podID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerID="2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944" exitCode=0 Oct 08 10:08:02 crc kubenswrapper[4744]: I1008 10:08:02.562886 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd7n4" event={"ID":"b76beff0-6871-4de0-a50c-82327bae3b2f","Type":"ContainerDied","Data":"2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944"} Oct 08 10:08:05 crc kubenswrapper[4744]: I1008 10:08:05.127185 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-scl94" podUID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerName="registry-server" probeResult="failure" output=< Oct 08 10:08:05 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 10:08:05 crc kubenswrapper[4744]: > Oct 08 10:08:09 crc kubenswrapper[4744]: E1008 10:08:09.188478 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944 is running failed: container process not found" containerID="2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 10:08:09 crc kubenswrapper[4744]: E1008 10:08:09.190638 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944 is running failed: container process not found" containerID="2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 10:08:09 crc kubenswrapper[4744]: E1008 10:08:09.191016 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944 is running failed: container process not found" containerID="2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 10:08:09 crc kubenswrapper[4744]: E1008 10:08:09.191066 4744 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-rd7n4" podUID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerName="registry-server" Oct 08 10:08:15 crc kubenswrapper[4744]: I1008 10:08:15.120884 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-scl94" podUID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerName="registry-server" probeResult="failure" output=< Oct 08 10:08:15 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 10:08:15 crc kubenswrapper[4744]: > Oct 08 10:08:19 crc kubenswrapper[4744]: E1008 10:08:19.184515 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944 is running failed: container process not found" containerID="2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 10:08:19 crc kubenswrapper[4744]: E1008 10:08:19.186954 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944 is running failed: container process not found" containerID="2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 10:08:19 crc kubenswrapper[4744]: E1008 10:08:19.187395 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944 is running failed: container process not found" containerID="2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944" cmd=["grpc_health_probe","-addr=:50051"] Oct 08 10:08:19 crc kubenswrapper[4744]: E1008 10:08:19.187531 4744 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-rd7n4" podUID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerName="registry-server" Oct 08 10:08:22 crc kubenswrapper[4744]: E1008 10:08:22.470694 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 08 10:08:22 crc kubenswrapper[4744]: E1008 10:08:22.483966 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h776v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(ba86d64c-cc8a-48b2-b70e-2ca6cf2145be): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 10:08:22 crc kubenswrapper[4744]: E1008 10:08:22.485815 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" Oct 08 10:08:22 crc kubenswrapper[4744]: E1008 10:08:22.825183 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.170613 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.211546 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b76beff0-6871-4de0-a50c-82327bae3b2f-catalog-content\") pod \"b76beff0-6871-4de0-a50c-82327bae3b2f\" (UID: \"b76beff0-6871-4de0-a50c-82327bae3b2f\") " Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.211712 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jchrh\" (UniqueName: \"kubernetes.io/projected/b76beff0-6871-4de0-a50c-82327bae3b2f-kube-api-access-jchrh\") pod \"b76beff0-6871-4de0-a50c-82327bae3b2f\" (UID: \"b76beff0-6871-4de0-a50c-82327bae3b2f\") " Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.211785 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b76beff0-6871-4de0-a50c-82327bae3b2f-utilities\") pod \"b76beff0-6871-4de0-a50c-82327bae3b2f\" (UID: \"b76beff0-6871-4de0-a50c-82327bae3b2f\") " Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.212503 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b76beff0-6871-4de0-a50c-82327bae3b2f-utilities" (OuterVolumeSpecName: "utilities") pod "b76beff0-6871-4de0-a50c-82327bae3b2f" (UID: "b76beff0-6871-4de0-a50c-82327bae3b2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.222957 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b76beff0-6871-4de0-a50c-82327bae3b2f-kube-api-access-jchrh" (OuterVolumeSpecName: "kube-api-access-jchrh") pod "b76beff0-6871-4de0-a50c-82327bae3b2f" (UID: "b76beff0-6871-4de0-a50c-82327bae3b2f"). InnerVolumeSpecName "kube-api-access-jchrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.246432 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b76beff0-6871-4de0-a50c-82327bae3b2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b76beff0-6871-4de0-a50c-82327bae3b2f" (UID: "b76beff0-6871-4de0-a50c-82327bae3b2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.314315 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b76beff0-6871-4de0-a50c-82327bae3b2f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.314408 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jchrh\" (UniqueName: \"kubernetes.io/projected/b76beff0-6871-4de0-a50c-82327bae3b2f-kube-api-access-jchrh\") on node \"crc\" DevicePath \"\"" Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.314430 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b76beff0-6871-4de0-a50c-82327bae3b2f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.859422 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rd7n4" event={"ID":"b76beff0-6871-4de0-a50c-82327bae3b2f","Type":"ContainerDied","Data":"2857add3449a6fa083596bf36239000821981c4b50433f6e6c69710c0ded6e20"} Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.859486 4744 scope.go:117] "RemoveContainer" containerID="2c81e39ead8c79cb524f10fe2da7b8c04863ab7b9f57e3692c27608e96a80944" Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.859673 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rd7n4" Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.894941 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rd7n4"] Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.903100 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rd7n4"] Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.910127 4744 scope.go:117] "RemoveContainer" containerID="cfab3e2f252a3a52d4b5067160ecb642e8301245ac6333211c98e130d7868697" Oct 08 10:08:23 crc kubenswrapper[4744]: I1008 10:08:23.945058 4744 scope.go:117] "RemoveContainer" containerID="d2f8a6f37511f36d941635a9fee729c7cb8a1affbecaf7fa32b3dc292cc3c4f1" Oct 08 10:08:24 crc kubenswrapper[4744]: I1008 10:08:24.124468 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:08:24 crc kubenswrapper[4744]: I1008 10:08:24.183175 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:08:25 crc kubenswrapper[4744]: I1008 10:08:25.416884 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-scl94"] Oct 08 10:08:25 crc kubenswrapper[4744]: I1008 10:08:25.464484 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b76beff0-6871-4de0-a50c-82327bae3b2f" path="/var/lib/kubelet/pods/b76beff0-6871-4de0-a50c-82327bae3b2f/volumes" Oct 08 10:08:25 crc kubenswrapper[4744]: I1008 10:08:25.881628 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-scl94" podUID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerName="registry-server" containerID="cri-o://9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39" gracePeriod=2 Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.393017 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.510386 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br9wz\" (UniqueName: \"kubernetes.io/projected/ee7f1dc6-9467-488a-9da6-947f88cb02ac-kube-api-access-br9wz\") pod \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\" (UID: \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\") " Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.510606 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee7f1dc6-9467-488a-9da6-947f88cb02ac-catalog-content\") pod \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\" (UID: \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\") " Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.510715 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee7f1dc6-9467-488a-9da6-947f88cb02ac-utilities\") pod \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\" (UID: \"ee7f1dc6-9467-488a-9da6-947f88cb02ac\") " Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.511922 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee7f1dc6-9467-488a-9da6-947f88cb02ac-utilities" (OuterVolumeSpecName: "utilities") pod "ee7f1dc6-9467-488a-9da6-947f88cb02ac" (UID: "ee7f1dc6-9467-488a-9da6-947f88cb02ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.522115 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee7f1dc6-9467-488a-9da6-947f88cb02ac-kube-api-access-br9wz" (OuterVolumeSpecName: "kube-api-access-br9wz") pod "ee7f1dc6-9467-488a-9da6-947f88cb02ac" (UID: "ee7f1dc6-9467-488a-9da6-947f88cb02ac"). InnerVolumeSpecName "kube-api-access-br9wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.591108 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee7f1dc6-9467-488a-9da6-947f88cb02ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee7f1dc6-9467-488a-9da6-947f88cb02ac" (UID: "ee7f1dc6-9467-488a-9da6-947f88cb02ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.613894 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee7f1dc6-9467-488a-9da6-947f88cb02ac-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.613934 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee7f1dc6-9467-488a-9da6-947f88cb02ac-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.613948 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br9wz\" (UniqueName: \"kubernetes.io/projected/ee7f1dc6-9467-488a-9da6-947f88cb02ac-kube-api-access-br9wz\") on node \"crc\" DevicePath \"\"" Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.890744 4744 generic.go:334] "Generic (PLEG): container finished" podID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerID="9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39" exitCode=0 Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.890798 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scl94" event={"ID":"ee7f1dc6-9467-488a-9da6-947f88cb02ac","Type":"ContainerDied","Data":"9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39"} Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.890813 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-scl94" Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.890839 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scl94" event={"ID":"ee7f1dc6-9467-488a-9da6-947f88cb02ac","Type":"ContainerDied","Data":"8f7b0b6a52e5f618fd51df66374c2243128917d1b680af9d73cac753ac6e8bf5"} Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.890867 4744 scope.go:117] "RemoveContainer" containerID="9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39" Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.916855 4744 scope.go:117] "RemoveContainer" containerID="b71b493ec4b83358c9f7e14a7a85dc52efc7eff769558b4b3e396b7ffe670bd8" Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.955090 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-scl94"] Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.965617 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-scl94"] Oct 08 10:08:26 crc kubenswrapper[4744]: I1008 10:08:26.968017 4744 scope.go:117] "RemoveContainer" containerID="870a934102c8ce02794d9d65f581e1d5f1df0176940d012e2427a4f3324a4212" Oct 08 10:08:27 crc kubenswrapper[4744]: I1008 10:08:27.028361 4744 scope.go:117] "RemoveContainer" containerID="9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39" Oct 08 10:08:27 crc kubenswrapper[4744]: E1008 10:08:27.028913 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39\": container with ID starting with 9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39 not found: ID does not exist" containerID="9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39" Oct 08 10:08:27 crc kubenswrapper[4744]: I1008 10:08:27.029040 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39"} err="failed to get container status \"9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39\": rpc error: code = NotFound desc = could not find container \"9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39\": container with ID starting with 9a5e1a534f30e723269a5832d3ac8d05129f6b1fb68c88927310463660a3ba39 not found: ID does not exist" Oct 08 10:08:27 crc kubenswrapper[4744]: I1008 10:08:27.029156 4744 scope.go:117] "RemoveContainer" containerID="b71b493ec4b83358c9f7e14a7a85dc52efc7eff769558b4b3e396b7ffe670bd8" Oct 08 10:08:27 crc kubenswrapper[4744]: E1008 10:08:27.029668 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b71b493ec4b83358c9f7e14a7a85dc52efc7eff769558b4b3e396b7ffe670bd8\": container with ID starting with b71b493ec4b83358c9f7e14a7a85dc52efc7eff769558b4b3e396b7ffe670bd8 not found: ID does not exist" containerID="b71b493ec4b83358c9f7e14a7a85dc52efc7eff769558b4b3e396b7ffe670bd8" Oct 08 10:08:27 crc kubenswrapper[4744]: I1008 10:08:27.029721 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b71b493ec4b83358c9f7e14a7a85dc52efc7eff769558b4b3e396b7ffe670bd8"} err="failed to get container status \"b71b493ec4b83358c9f7e14a7a85dc52efc7eff769558b4b3e396b7ffe670bd8\": rpc error: code = NotFound desc = could not find container \"b71b493ec4b83358c9f7e14a7a85dc52efc7eff769558b4b3e396b7ffe670bd8\": container with ID starting with b71b493ec4b83358c9f7e14a7a85dc52efc7eff769558b4b3e396b7ffe670bd8 not found: ID does not exist" Oct 08 10:08:27 crc kubenswrapper[4744]: I1008 10:08:27.029753 4744 scope.go:117] "RemoveContainer" containerID="870a934102c8ce02794d9d65f581e1d5f1df0176940d012e2427a4f3324a4212" Oct 08 10:08:27 crc kubenswrapper[4744]: E1008 10:08:27.030571 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"870a934102c8ce02794d9d65f581e1d5f1df0176940d012e2427a4f3324a4212\": container with ID starting with 870a934102c8ce02794d9d65f581e1d5f1df0176940d012e2427a4f3324a4212 not found: ID does not exist" containerID="870a934102c8ce02794d9d65f581e1d5f1df0176940d012e2427a4f3324a4212" Oct 08 10:08:27 crc kubenswrapper[4744]: I1008 10:08:27.030665 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"870a934102c8ce02794d9d65f581e1d5f1df0176940d012e2427a4f3324a4212"} err="failed to get container status \"870a934102c8ce02794d9d65f581e1d5f1df0176940d012e2427a4f3324a4212\": rpc error: code = NotFound desc = could not find container \"870a934102c8ce02794d9d65f581e1d5f1df0176940d012e2427a4f3324a4212\": container with ID starting with 870a934102c8ce02794d9d65f581e1d5f1df0176940d012e2427a4f3324a4212 not found: ID does not exist" Oct 08 10:08:27 crc kubenswrapper[4744]: I1008 10:08:27.466006 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" path="/var/lib/kubelet/pods/ee7f1dc6-9467-488a-9da6-947f88cb02ac/volumes" Oct 08 10:08:38 crc kubenswrapper[4744]: I1008 10:08:38.963208 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 08 10:08:40 crc kubenswrapper[4744]: I1008 10:08:40.037329 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be","Type":"ContainerStarted","Data":"b7c39adb1c4c22a29d29fad2de16bf87f1c7cef29adf3e343f39ee82be31941f"} Oct 08 10:08:40 crc kubenswrapper[4744]: I1008 10:08:40.058434 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.428642538 podStartE2EDuration="1m4.058416244s" podCreationTimestamp="2025-10-08 10:07:36 +0000 UTC" firstStartedPulling="2025-10-08 10:07:38.330763083 +0000 UTC m=+3353.578408322" lastFinishedPulling="2025-10-08 10:08:38.960536789 +0000 UTC m=+3414.208182028" observedRunningTime="2025-10-08 10:08:40.054698782 +0000 UTC m=+3415.302344081" watchObservedRunningTime="2025-10-08 10:08:40.058416244 +0000 UTC m=+3415.306061483" Oct 08 10:09:19 crc kubenswrapper[4744]: I1008 10:09:19.689953 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:09:19 crc kubenswrapper[4744]: I1008 10:09:19.691461 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:09:49 crc kubenswrapper[4744]: I1008 10:09:49.690233 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:09:49 crc kubenswrapper[4744]: I1008 10:09:49.690763 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:10:19 crc kubenswrapper[4744]: I1008 10:10:19.691229 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:10:19 crc kubenswrapper[4744]: I1008 10:10:19.692394 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:10:19 crc kubenswrapper[4744]: I1008 10:10:19.692478 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 10:10:19 crc kubenswrapper[4744]: I1008 10:10:19.693817 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"24813bdc7353e22419b64c64e6f4e18c33e77fd646809a089b4e5a14586c42b3"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 10:10:19 crc kubenswrapper[4744]: I1008 10:10:19.693915 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://24813bdc7353e22419b64c64e6f4e18c33e77fd646809a089b4e5a14586c42b3" gracePeriod=600 Oct 08 10:10:20 crc kubenswrapper[4744]: I1008 10:10:20.074045 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="24813bdc7353e22419b64c64e6f4e18c33e77fd646809a089b4e5a14586c42b3" exitCode=0 Oct 08 10:10:20 crc kubenswrapper[4744]: I1008 10:10:20.074098 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"24813bdc7353e22419b64c64e6f4e18c33e77fd646809a089b4e5a14586c42b3"} Oct 08 10:10:20 crc kubenswrapper[4744]: I1008 10:10:20.074344 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a"} Oct 08 10:10:20 crc kubenswrapper[4744]: I1008 10:10:20.074381 4744 scope.go:117] "RemoveContainer" containerID="d1855adba97c87f46558ca0a25de4b70ece9d10ee8da73430eab6da68b734bf8" Oct 08 10:12:49 crc kubenswrapper[4744]: I1008 10:12:49.690910 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:12:49 crc kubenswrapper[4744]: I1008 10:12:49.691473 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:13:19 crc kubenswrapper[4744]: I1008 10:13:19.690969 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:13:19 crc kubenswrapper[4744]: I1008 10:13:19.692540 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:13:49 crc kubenswrapper[4744]: I1008 10:13:49.690202 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:13:49 crc kubenswrapper[4744]: I1008 10:13:49.690776 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:13:49 crc kubenswrapper[4744]: I1008 10:13:49.690827 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 10:13:49 crc kubenswrapper[4744]: I1008 10:13:49.691706 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 10:13:49 crc kubenswrapper[4744]: I1008 10:13:49.691772 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" gracePeriod=600 Oct 08 10:13:49 crc kubenswrapper[4744]: E1008 10:13:49.811242 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:13:50 crc kubenswrapper[4744]: I1008 10:13:50.054059 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" exitCode=0 Oct 08 10:13:50 crc kubenswrapper[4744]: I1008 10:13:50.054102 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a"} Oct 08 10:13:50 crc kubenswrapper[4744]: I1008 10:13:50.054136 4744 scope.go:117] "RemoveContainer" containerID="24813bdc7353e22419b64c64e6f4e18c33e77fd646809a089b4e5a14586c42b3" Oct 08 10:13:50 crc kubenswrapper[4744]: I1008 10:13:50.054880 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:13:50 crc kubenswrapper[4744]: E1008 10:13:50.055118 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:14:04 crc kubenswrapper[4744]: I1008 10:14:04.453857 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:14:04 crc kubenswrapper[4744]: E1008 10:14:04.454544 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:14:19 crc kubenswrapper[4744]: I1008 10:14:19.457972 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:14:19 crc kubenswrapper[4744]: E1008 10:14:19.458655 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:14:34 crc kubenswrapper[4744]: I1008 10:14:34.455288 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:14:34 crc kubenswrapper[4744]: E1008 10:14:34.457293 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:14:47 crc kubenswrapper[4744]: I1008 10:14:47.453983 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:14:47 crc kubenswrapper[4744]: E1008 10:14:47.454744 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.126218 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-v7c88"] Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.145903 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-v7c88"] Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.178902 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf"] Oct 08 10:15:00 crc kubenswrapper[4744]: E1008 10:15:00.179316 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerName="registry-server" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.179335 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerName="registry-server" Oct 08 10:15:00 crc kubenswrapper[4744]: E1008 10:15:00.179357 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerName="registry-server" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.179383 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerName="registry-server" Oct 08 10:15:00 crc kubenswrapper[4744]: E1008 10:15:00.179401 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerName="extract-content" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.179408 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerName="extract-content" Oct 08 10:15:00 crc kubenswrapper[4744]: E1008 10:15:00.179422 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerName="extract-utilities" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.179428 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerName="extract-utilities" Oct 08 10:15:00 crc kubenswrapper[4744]: E1008 10:15:00.179436 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerName="extract-utilities" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.179442 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerName="extract-utilities" Oct 08 10:15:00 crc kubenswrapper[4744]: E1008 10:15:00.179474 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerName="extract-content" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.179479 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerName="extract-content" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.179646 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7f1dc6-9467-488a-9da6-947f88cb02ac" containerName="registry-server" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.179674 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b76beff0-6871-4de0-a50c-82327bae3b2f" containerName="registry-server" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.182742 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.184710 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.185116 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.199916 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf"] Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.279923 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-config-volume\") pod \"collect-profiles-29331975-55vrf\" (UID: \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.280104 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9pqb\" (UniqueName: \"kubernetes.io/projected/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-kube-api-access-l9pqb\") pod \"collect-profiles-29331975-55vrf\" (UID: \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.280130 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-secret-volume\") pod \"collect-profiles-29331975-55vrf\" (UID: \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.382237 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9pqb\" (UniqueName: \"kubernetes.io/projected/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-kube-api-access-l9pqb\") pod \"collect-profiles-29331975-55vrf\" (UID: \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.382311 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-secret-volume\") pod \"collect-profiles-29331975-55vrf\" (UID: \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.382518 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-config-volume\") pod \"collect-profiles-29331975-55vrf\" (UID: \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.383601 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-config-volume\") pod \"collect-profiles-29331975-55vrf\" (UID: \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.393210 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-secret-volume\") pod \"collect-profiles-29331975-55vrf\" (UID: \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.398891 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9pqb\" (UniqueName: \"kubernetes.io/projected/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-kube-api-access-l9pqb\") pod \"collect-profiles-29331975-55vrf\" (UID: \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.454480 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:15:00 crc kubenswrapper[4744]: E1008 10:15:00.454782 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:15:00 crc kubenswrapper[4744]: I1008 10:15:00.504933 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:01 crc kubenswrapper[4744]: I1008 10:15:01.096852 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf"] Oct 08 10:15:01 crc kubenswrapper[4744]: I1008 10:15:01.465909 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="119d6cdd-c521-4ad3-b4c0-ceb7eb956281" path="/var/lib/kubelet/pods/119d6cdd-c521-4ad3-b4c0-ceb7eb956281/volumes" Oct 08 10:15:01 crc kubenswrapper[4744]: I1008 10:15:01.676291 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" event={"ID":"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d","Type":"ContainerStarted","Data":"fb799315482acb249a54e47eb4403a1fc18955d033f9b62809863dd7610f1180"} Oct 08 10:15:01 crc kubenswrapper[4744]: I1008 10:15:01.676348 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" event={"ID":"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d","Type":"ContainerStarted","Data":"ec6eea1a169e34ae4ed76c0e6ab4c53a030bc2e15e054d1d6b94c7efde0dd9e4"} Oct 08 10:15:02 crc kubenswrapper[4744]: I1008 10:15:02.688575 4744 generic.go:334] "Generic (PLEG): container finished" podID="c7dc14cb-458b-4cbe-8467-b2c2e2c7703d" containerID="fb799315482acb249a54e47eb4403a1fc18955d033f9b62809863dd7610f1180" exitCode=0 Oct 08 10:15:02 crc kubenswrapper[4744]: I1008 10:15:02.688645 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" event={"ID":"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d","Type":"ContainerDied","Data":"fb799315482acb249a54e47eb4403a1fc18955d033f9b62809863dd7610f1180"} Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.157603 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.271217 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-secret-volume\") pod \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\" (UID: \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\") " Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.271819 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9pqb\" (UniqueName: \"kubernetes.io/projected/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-kube-api-access-l9pqb\") pod \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\" (UID: \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\") " Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.271872 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-config-volume\") pod \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\" (UID: \"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d\") " Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.273166 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-config-volume" (OuterVolumeSpecName: "config-volume") pod "c7dc14cb-458b-4cbe-8467-b2c2e2c7703d" (UID: "c7dc14cb-458b-4cbe-8467-b2c2e2c7703d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.277578 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c7dc14cb-458b-4cbe-8467-b2c2e2c7703d" (UID: "c7dc14cb-458b-4cbe-8467-b2c2e2c7703d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.277739 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-kube-api-access-l9pqb" (OuterVolumeSpecName: "kube-api-access-l9pqb") pod "c7dc14cb-458b-4cbe-8467-b2c2e2c7703d" (UID: "c7dc14cb-458b-4cbe-8467-b2c2e2c7703d"). InnerVolumeSpecName "kube-api-access-l9pqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.374677 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.374710 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9pqb\" (UniqueName: \"kubernetes.io/projected/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-kube-api-access-l9pqb\") on node \"crc\" DevicePath \"\"" Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.374720 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7dc14cb-458b-4cbe-8467-b2c2e2c7703d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.710982 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" event={"ID":"c7dc14cb-458b-4cbe-8467-b2c2e2c7703d","Type":"ContainerDied","Data":"ec6eea1a169e34ae4ed76c0e6ab4c53a030bc2e15e054d1d6b94c7efde0dd9e4"} Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.711029 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec6eea1a169e34ae4ed76c0e6ab4c53a030bc2e15e054d1d6b94c7efde0dd9e4" Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.711446 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331975-55vrf" Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.796464 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5"] Oct 08 10:15:04 crc kubenswrapper[4744]: I1008 10:15:04.805770 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331930-zgwg5"] Oct 08 10:15:05 crc kubenswrapper[4744]: I1008 10:15:05.473074 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="162601ad-6234-499e-acfe-6e8fba25d8a0" path="/var/lib/kubelet/pods/162601ad-6234-499e-acfe-6e8fba25d8a0/volumes" Oct 08 10:15:11 crc kubenswrapper[4744]: I1008 10:15:11.453359 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:15:11 crc kubenswrapper[4744]: E1008 10:15:11.453974 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:15:19 crc kubenswrapper[4744]: I1008 10:15:19.028938 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-5c7c-account-create-ww5bj"] Oct 08 10:15:19 crc kubenswrapper[4744]: I1008 10:15:19.036976 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-5c7c-account-create-ww5bj"] Oct 08 10:15:19 crc kubenswrapper[4744]: I1008 10:15:19.467730 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308c5a41-6d01-4684-9ab0-a23e0e372252" path="/var/lib/kubelet/pods/308c5a41-6d01-4684-9ab0-a23e0e372252/volumes" Oct 08 10:15:20 crc kubenswrapper[4744]: I1008 10:15:20.959974 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hmcbb"] Oct 08 10:15:20 crc kubenswrapper[4744]: E1008 10:15:20.960950 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7dc14cb-458b-4cbe-8467-b2c2e2c7703d" containerName="collect-profiles" Oct 08 10:15:20 crc kubenswrapper[4744]: I1008 10:15:20.961190 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7dc14cb-458b-4cbe-8467-b2c2e2c7703d" containerName="collect-profiles" Oct 08 10:15:20 crc kubenswrapper[4744]: I1008 10:15:20.962621 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7dc14cb-458b-4cbe-8467-b2c2e2c7703d" containerName="collect-profiles" Oct 08 10:15:20 crc kubenswrapper[4744]: I1008 10:15:20.964322 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:21 crc kubenswrapper[4744]: I1008 10:15:21.007434 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hmcbb"] Oct 08 10:15:21 crc kubenswrapper[4744]: I1008 10:15:21.064811 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bac70c8-4f11-4834-9958-90c2aec7787f-catalog-content\") pod \"certified-operators-hmcbb\" (UID: \"7bac70c8-4f11-4834-9958-90c2aec7787f\") " pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:21 crc kubenswrapper[4744]: I1008 10:15:21.064990 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bac70c8-4f11-4834-9958-90c2aec7787f-utilities\") pod \"certified-operators-hmcbb\" (UID: \"7bac70c8-4f11-4834-9958-90c2aec7787f\") " pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:21 crc kubenswrapper[4744]: I1008 10:15:21.065041 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49q4s\" (UniqueName: \"kubernetes.io/projected/7bac70c8-4f11-4834-9958-90c2aec7787f-kube-api-access-49q4s\") pod \"certified-operators-hmcbb\" (UID: \"7bac70c8-4f11-4834-9958-90c2aec7787f\") " pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:21 crc kubenswrapper[4744]: I1008 10:15:21.167942 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bac70c8-4f11-4834-9958-90c2aec7787f-catalog-content\") pod \"certified-operators-hmcbb\" (UID: \"7bac70c8-4f11-4834-9958-90c2aec7787f\") " pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:21 crc kubenswrapper[4744]: I1008 10:15:21.168008 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bac70c8-4f11-4834-9958-90c2aec7787f-utilities\") pod \"certified-operators-hmcbb\" (UID: \"7bac70c8-4f11-4834-9958-90c2aec7787f\") " pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:21 crc kubenswrapper[4744]: I1008 10:15:21.168034 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49q4s\" (UniqueName: \"kubernetes.io/projected/7bac70c8-4f11-4834-9958-90c2aec7787f-kube-api-access-49q4s\") pod \"certified-operators-hmcbb\" (UID: \"7bac70c8-4f11-4834-9958-90c2aec7787f\") " pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:21 crc kubenswrapper[4744]: I1008 10:15:21.169962 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bac70c8-4f11-4834-9958-90c2aec7787f-catalog-content\") pod \"certified-operators-hmcbb\" (UID: \"7bac70c8-4f11-4834-9958-90c2aec7787f\") " pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:21 crc kubenswrapper[4744]: I1008 10:15:21.169980 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bac70c8-4f11-4834-9958-90c2aec7787f-utilities\") pod \"certified-operators-hmcbb\" (UID: \"7bac70c8-4f11-4834-9958-90c2aec7787f\") " pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:21 crc kubenswrapper[4744]: I1008 10:15:21.194244 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49q4s\" (UniqueName: \"kubernetes.io/projected/7bac70c8-4f11-4834-9958-90c2aec7787f-kube-api-access-49q4s\") pod \"certified-operators-hmcbb\" (UID: \"7bac70c8-4f11-4834-9958-90c2aec7787f\") " pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:21 crc kubenswrapper[4744]: I1008 10:15:21.284074 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:21 crc kubenswrapper[4744]: I1008 10:15:21.915104 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hmcbb"] Oct 08 10:15:22 crc kubenswrapper[4744]: I1008 10:15:22.453987 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:15:22 crc kubenswrapper[4744]: E1008 10:15:22.454585 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:15:22 crc kubenswrapper[4744]: I1008 10:15:22.887468 4744 generic.go:334] "Generic (PLEG): container finished" podID="7bac70c8-4f11-4834-9958-90c2aec7787f" containerID="2c356de25e2c45251e0fe20e02cf87f3e92239721a09644f026a68994b9d2136" exitCode=0 Oct 08 10:15:22 crc kubenswrapper[4744]: I1008 10:15:22.887661 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmcbb" event={"ID":"7bac70c8-4f11-4834-9958-90c2aec7787f","Type":"ContainerDied","Data":"2c356de25e2c45251e0fe20e02cf87f3e92239721a09644f026a68994b9d2136"} Oct 08 10:15:22 crc kubenswrapper[4744]: I1008 10:15:22.888559 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmcbb" event={"ID":"7bac70c8-4f11-4834-9958-90c2aec7787f","Type":"ContainerStarted","Data":"083b4f69d4638544e889ff32c64be6fc52a2c9e2ab1a3f5fe15d10685ed9b1a1"} Oct 08 10:15:22 crc kubenswrapper[4744]: I1008 10:15:22.891147 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.359437 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m8smv"] Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.366128 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.371854 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m8smv"] Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.518387 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlvcn\" (UniqueName: \"kubernetes.io/projected/97c071e6-18f7-4a47-9cb1-e74ce7084109-kube-api-access-dlvcn\") pod \"community-operators-m8smv\" (UID: \"97c071e6-18f7-4a47-9cb1-e74ce7084109\") " pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.518510 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c071e6-18f7-4a47-9cb1-e74ce7084109-catalog-content\") pod \"community-operators-m8smv\" (UID: \"97c071e6-18f7-4a47-9cb1-e74ce7084109\") " pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.518680 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c071e6-18f7-4a47-9cb1-e74ce7084109-utilities\") pod \"community-operators-m8smv\" (UID: \"97c071e6-18f7-4a47-9cb1-e74ce7084109\") " pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.621135 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlvcn\" (UniqueName: \"kubernetes.io/projected/97c071e6-18f7-4a47-9cb1-e74ce7084109-kube-api-access-dlvcn\") pod \"community-operators-m8smv\" (UID: \"97c071e6-18f7-4a47-9cb1-e74ce7084109\") " pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.621263 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c071e6-18f7-4a47-9cb1-e74ce7084109-catalog-content\") pod \"community-operators-m8smv\" (UID: \"97c071e6-18f7-4a47-9cb1-e74ce7084109\") " pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.621309 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c071e6-18f7-4a47-9cb1-e74ce7084109-utilities\") pod \"community-operators-m8smv\" (UID: \"97c071e6-18f7-4a47-9cb1-e74ce7084109\") " pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.621836 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c071e6-18f7-4a47-9cb1-e74ce7084109-utilities\") pod \"community-operators-m8smv\" (UID: \"97c071e6-18f7-4a47-9cb1-e74ce7084109\") " pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.621896 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c071e6-18f7-4a47-9cb1-e74ce7084109-catalog-content\") pod \"community-operators-m8smv\" (UID: \"97c071e6-18f7-4a47-9cb1-e74ce7084109\") " pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.641083 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlvcn\" (UniqueName: \"kubernetes.io/projected/97c071e6-18f7-4a47-9cb1-e74ce7084109-kube-api-access-dlvcn\") pod \"community-operators-m8smv\" (UID: \"97c071e6-18f7-4a47-9cb1-e74ce7084109\") " pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:23 crc kubenswrapper[4744]: I1008 10:15:23.709645 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:24 crc kubenswrapper[4744]: I1008 10:15:24.300086 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m8smv"] Oct 08 10:15:24 crc kubenswrapper[4744]: E1008 10:15:24.661980 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97c071e6_18f7_4a47_9cb1_e74ce7084109.slice/crio-3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97c071e6_18f7_4a47_9cb1_e74ce7084109.slice/crio-conmon-3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56.scope\": RecentStats: unable to find data in memory cache]" Oct 08 10:15:24 crc kubenswrapper[4744]: I1008 10:15:24.916005 4744 generic.go:334] "Generic (PLEG): container finished" podID="97c071e6-18f7-4a47-9cb1-e74ce7084109" containerID="3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56" exitCode=0 Oct 08 10:15:24 crc kubenswrapper[4744]: I1008 10:15:24.916069 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m8smv" event={"ID":"97c071e6-18f7-4a47-9cb1-e74ce7084109","Type":"ContainerDied","Data":"3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56"} Oct 08 10:15:24 crc kubenswrapper[4744]: I1008 10:15:24.916093 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m8smv" event={"ID":"97c071e6-18f7-4a47-9cb1-e74ce7084109","Type":"ContainerStarted","Data":"b008a8502b2b6dfaac9fd89fc5aa7541a5e55870460d968febf59833c0979bd7"} Oct 08 10:15:24 crc kubenswrapper[4744]: I1008 10:15:24.919474 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmcbb" event={"ID":"7bac70c8-4f11-4834-9958-90c2aec7787f","Type":"ContainerStarted","Data":"31e8716219f628831671b6d8113c62262b9bb47eb486bd546e4ee47d914e2342"} Oct 08 10:15:26 crc kubenswrapper[4744]: I1008 10:15:26.957081 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m8smv" event={"ID":"97c071e6-18f7-4a47-9cb1-e74ce7084109","Type":"ContainerStarted","Data":"1052d7d8b79a555fb4282f3560936eb1292d7b95c05cbb4ce2332d2af418f118"} Oct 08 10:15:27 crc kubenswrapper[4744]: I1008 10:15:27.969379 4744 generic.go:334] "Generic (PLEG): container finished" podID="7bac70c8-4f11-4834-9958-90c2aec7787f" containerID="31e8716219f628831671b6d8113c62262b9bb47eb486bd546e4ee47d914e2342" exitCode=0 Oct 08 10:15:27 crc kubenswrapper[4744]: I1008 10:15:27.969400 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmcbb" event={"ID":"7bac70c8-4f11-4834-9958-90c2aec7787f","Type":"ContainerDied","Data":"31e8716219f628831671b6d8113c62262b9bb47eb486bd546e4ee47d914e2342"} Oct 08 10:15:30 crc kubenswrapper[4744]: I1008 10:15:30.005265 4744 generic.go:334] "Generic (PLEG): container finished" podID="97c071e6-18f7-4a47-9cb1-e74ce7084109" containerID="1052d7d8b79a555fb4282f3560936eb1292d7b95c05cbb4ce2332d2af418f118" exitCode=0 Oct 08 10:15:30 crc kubenswrapper[4744]: I1008 10:15:30.005532 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m8smv" event={"ID":"97c071e6-18f7-4a47-9cb1-e74ce7084109","Type":"ContainerDied","Data":"1052d7d8b79a555fb4282f3560936eb1292d7b95c05cbb4ce2332d2af418f118"} Oct 08 10:15:30 crc kubenswrapper[4744]: I1008 10:15:30.011952 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmcbb" event={"ID":"7bac70c8-4f11-4834-9958-90c2aec7787f","Type":"ContainerStarted","Data":"1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa"} Oct 08 10:15:30 crc kubenswrapper[4744]: I1008 10:15:30.067830 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hmcbb" podStartSLOduration=4.186194925 podStartE2EDuration="10.067813389s" podCreationTimestamp="2025-10-08 10:15:20 +0000 UTC" firstStartedPulling="2025-10-08 10:15:22.890745188 +0000 UTC m=+3818.138390467" lastFinishedPulling="2025-10-08 10:15:28.772363692 +0000 UTC m=+3824.020008931" observedRunningTime="2025-10-08 10:15:30.062910503 +0000 UTC m=+3825.310555752" watchObservedRunningTime="2025-10-08 10:15:30.067813389 +0000 UTC m=+3825.315458628" Oct 08 10:15:31 crc kubenswrapper[4744]: I1008 10:15:31.285244 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:31 crc kubenswrapper[4744]: I1008 10:15:31.285565 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:32 crc kubenswrapper[4744]: I1008 10:15:32.044024 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m8smv" event={"ID":"97c071e6-18f7-4a47-9cb1-e74ce7084109","Type":"ContainerStarted","Data":"05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6"} Oct 08 10:15:32 crc kubenswrapper[4744]: I1008 10:15:32.067051 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m8smv" podStartSLOduration=3.162054524 podStartE2EDuration="9.067025143s" podCreationTimestamp="2025-10-08 10:15:23 +0000 UTC" firstStartedPulling="2025-10-08 10:15:24.917920167 +0000 UTC m=+3820.165565406" lastFinishedPulling="2025-10-08 10:15:30.822890786 +0000 UTC m=+3826.070536025" observedRunningTime="2025-10-08 10:15:32.059159956 +0000 UTC m=+3827.306805205" watchObservedRunningTime="2025-10-08 10:15:32.067025143 +0000 UTC m=+3827.314670372" Oct 08 10:15:32 crc kubenswrapper[4744]: I1008 10:15:32.330887 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-hmcbb" podUID="7bac70c8-4f11-4834-9958-90c2aec7787f" containerName="registry-server" probeResult="failure" output=< Oct 08 10:15:32 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 10:15:32 crc kubenswrapper[4744]: > Oct 08 10:15:33 crc kubenswrapper[4744]: I1008 10:15:33.710408 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:33 crc kubenswrapper[4744]: I1008 10:15:33.710781 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:34 crc kubenswrapper[4744]: I1008 10:15:34.813939 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-m8smv" podUID="97c071e6-18f7-4a47-9cb1-e74ce7084109" containerName="registry-server" probeResult="failure" output=< Oct 08 10:15:34 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 10:15:34 crc kubenswrapper[4744]: > Oct 08 10:15:37 crc kubenswrapper[4744]: I1008 10:15:37.454679 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:15:37 crc kubenswrapper[4744]: E1008 10:15:37.455282 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:15:41 crc kubenswrapper[4744]: I1008 10:15:41.378880 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:41 crc kubenswrapper[4744]: I1008 10:15:41.450493 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:41 crc kubenswrapper[4744]: I1008 10:15:41.618066 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hmcbb"] Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.142767 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hmcbb" podUID="7bac70c8-4f11-4834-9958-90c2aec7787f" containerName="registry-server" containerID="cri-o://1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa" gracePeriod=2 Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.766690 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.788506 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.852146 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.853219 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49q4s\" (UniqueName: \"kubernetes.io/projected/7bac70c8-4f11-4834-9958-90c2aec7787f-kube-api-access-49q4s\") pod \"7bac70c8-4f11-4834-9958-90c2aec7787f\" (UID: \"7bac70c8-4f11-4834-9958-90c2aec7787f\") " Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.853546 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bac70c8-4f11-4834-9958-90c2aec7787f-catalog-content\") pod \"7bac70c8-4f11-4834-9958-90c2aec7787f\" (UID: \"7bac70c8-4f11-4834-9958-90c2aec7787f\") " Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.853887 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bac70c8-4f11-4834-9958-90c2aec7787f-utilities\") pod \"7bac70c8-4f11-4834-9958-90c2aec7787f\" (UID: \"7bac70c8-4f11-4834-9958-90c2aec7787f\") " Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.855277 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bac70c8-4f11-4834-9958-90c2aec7787f-utilities" (OuterVolumeSpecName: "utilities") pod "7bac70c8-4f11-4834-9958-90c2aec7787f" (UID: "7bac70c8-4f11-4834-9958-90c2aec7787f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.881851 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bac70c8-4f11-4834-9958-90c2aec7787f-kube-api-access-49q4s" (OuterVolumeSpecName: "kube-api-access-49q4s") pod "7bac70c8-4f11-4834-9958-90c2aec7787f" (UID: "7bac70c8-4f11-4834-9958-90c2aec7787f"). InnerVolumeSpecName "kube-api-access-49q4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.913891 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bac70c8-4f11-4834-9958-90c2aec7787f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7bac70c8-4f11-4834-9958-90c2aec7787f" (UID: "7bac70c8-4f11-4834-9958-90c2aec7787f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.957558 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bac70c8-4f11-4834-9958-90c2aec7787f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.957594 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49q4s\" (UniqueName: \"kubernetes.io/projected/7bac70c8-4f11-4834-9958-90c2aec7787f-kube-api-access-49q4s\") on node \"crc\" DevicePath \"\"" Oct 08 10:15:43 crc kubenswrapper[4744]: I1008 10:15:43.957605 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bac70c8-4f11-4834-9958-90c2aec7787f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.169078 4744 generic.go:334] "Generic (PLEG): container finished" podID="7bac70c8-4f11-4834-9958-90c2aec7787f" containerID="1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa" exitCode=0 Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.169151 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hmcbb" Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.169213 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmcbb" event={"ID":"7bac70c8-4f11-4834-9958-90c2aec7787f","Type":"ContainerDied","Data":"1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa"} Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.169238 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmcbb" event={"ID":"7bac70c8-4f11-4834-9958-90c2aec7787f","Type":"ContainerDied","Data":"083b4f69d4638544e889ff32c64be6fc52a2c9e2ab1a3f5fe15d10685ed9b1a1"} Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.169270 4744 scope.go:117] "RemoveContainer" containerID="1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa" Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.206712 4744 scope.go:117] "RemoveContainer" containerID="31e8716219f628831671b6d8113c62262b9bb47eb486bd546e4ee47d914e2342" Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.213012 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hmcbb"] Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.222119 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hmcbb"] Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.230567 4744 scope.go:117] "RemoveContainer" containerID="2c356de25e2c45251e0fe20e02cf87f3e92239721a09644f026a68994b9d2136" Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.286734 4744 scope.go:117] "RemoveContainer" containerID="1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa" Oct 08 10:15:44 crc kubenswrapper[4744]: E1008 10:15:44.287357 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa\": container with ID starting with 1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa not found: ID does not exist" containerID="1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa" Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.287447 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa"} err="failed to get container status \"1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa\": rpc error: code = NotFound desc = could not find container \"1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa\": container with ID starting with 1e8e686af754b98adb7605eeae251d22b8f476cad8015b7e5b6c972585bc37fa not found: ID does not exist" Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.287485 4744 scope.go:117] "RemoveContainer" containerID="31e8716219f628831671b6d8113c62262b9bb47eb486bd546e4ee47d914e2342" Oct 08 10:15:44 crc kubenswrapper[4744]: E1008 10:15:44.287857 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31e8716219f628831671b6d8113c62262b9bb47eb486bd546e4ee47d914e2342\": container with ID starting with 31e8716219f628831671b6d8113c62262b9bb47eb486bd546e4ee47d914e2342 not found: ID does not exist" containerID="31e8716219f628831671b6d8113c62262b9bb47eb486bd546e4ee47d914e2342" Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.287892 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31e8716219f628831671b6d8113c62262b9bb47eb486bd546e4ee47d914e2342"} err="failed to get container status \"31e8716219f628831671b6d8113c62262b9bb47eb486bd546e4ee47d914e2342\": rpc error: code = NotFound desc = could not find container \"31e8716219f628831671b6d8113c62262b9bb47eb486bd546e4ee47d914e2342\": container with ID starting with 31e8716219f628831671b6d8113c62262b9bb47eb486bd546e4ee47d914e2342 not found: ID does not exist" Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.287920 4744 scope.go:117] "RemoveContainer" containerID="2c356de25e2c45251e0fe20e02cf87f3e92239721a09644f026a68994b9d2136" Oct 08 10:15:44 crc kubenswrapper[4744]: E1008 10:15:44.288647 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c356de25e2c45251e0fe20e02cf87f3e92239721a09644f026a68994b9d2136\": container with ID starting with 2c356de25e2c45251e0fe20e02cf87f3e92239721a09644f026a68994b9d2136 not found: ID does not exist" containerID="2c356de25e2c45251e0fe20e02cf87f3e92239721a09644f026a68994b9d2136" Oct 08 10:15:44 crc kubenswrapper[4744]: I1008 10:15:44.288694 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c356de25e2c45251e0fe20e02cf87f3e92239721a09644f026a68994b9d2136"} err="failed to get container status \"2c356de25e2c45251e0fe20e02cf87f3e92239721a09644f026a68994b9d2136\": rpc error: code = NotFound desc = could not find container \"2c356de25e2c45251e0fe20e02cf87f3e92239721a09644f026a68994b9d2136\": container with ID starting with 2c356de25e2c45251e0fe20e02cf87f3e92239721a09644f026a68994b9d2136 not found: ID does not exist" Oct 08 10:15:45 crc kubenswrapper[4744]: I1008 10:15:45.052401 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-s82nq"] Oct 08 10:15:45 crc kubenswrapper[4744]: I1008 10:15:45.061832 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-s82nq"] Oct 08 10:15:45 crc kubenswrapper[4744]: I1008 10:15:45.474724 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bac70c8-4f11-4834-9958-90c2aec7787f" path="/var/lib/kubelet/pods/7bac70c8-4f11-4834-9958-90c2aec7787f/volumes" Oct 08 10:15:45 crc kubenswrapper[4744]: I1008 10:15:45.476630 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="975dbb77-5e8d-4192-885d-57910d229671" path="/var/lib/kubelet/pods/975dbb77-5e8d-4192-885d-57910d229671/volumes" Oct 08 10:15:46 crc kubenswrapper[4744]: I1008 10:15:46.230355 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m8smv"] Oct 08 10:15:46 crc kubenswrapper[4744]: I1008 10:15:46.231201 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m8smv" podUID="97c071e6-18f7-4a47-9cb1-e74ce7084109" containerName="registry-server" containerID="cri-o://05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6" gracePeriod=2 Oct 08 10:15:46 crc kubenswrapper[4744]: I1008 10:15:46.770546 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:46 crc kubenswrapper[4744]: I1008 10:15:46.825672 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c071e6-18f7-4a47-9cb1-e74ce7084109-catalog-content\") pod \"97c071e6-18f7-4a47-9cb1-e74ce7084109\" (UID: \"97c071e6-18f7-4a47-9cb1-e74ce7084109\") " Oct 08 10:15:46 crc kubenswrapper[4744]: I1008 10:15:46.825749 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c071e6-18f7-4a47-9cb1-e74ce7084109-utilities\") pod \"97c071e6-18f7-4a47-9cb1-e74ce7084109\" (UID: \"97c071e6-18f7-4a47-9cb1-e74ce7084109\") " Oct 08 10:15:46 crc kubenswrapper[4744]: I1008 10:15:46.825807 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlvcn\" (UniqueName: \"kubernetes.io/projected/97c071e6-18f7-4a47-9cb1-e74ce7084109-kube-api-access-dlvcn\") pod \"97c071e6-18f7-4a47-9cb1-e74ce7084109\" (UID: \"97c071e6-18f7-4a47-9cb1-e74ce7084109\") " Oct 08 10:15:46 crc kubenswrapper[4744]: I1008 10:15:46.826536 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97c071e6-18f7-4a47-9cb1-e74ce7084109-utilities" (OuterVolumeSpecName: "utilities") pod "97c071e6-18f7-4a47-9cb1-e74ce7084109" (UID: "97c071e6-18f7-4a47-9cb1-e74ce7084109"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:15:46 crc kubenswrapper[4744]: I1008 10:15:46.833284 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97c071e6-18f7-4a47-9cb1-e74ce7084109-kube-api-access-dlvcn" (OuterVolumeSpecName: "kube-api-access-dlvcn") pod "97c071e6-18f7-4a47-9cb1-e74ce7084109" (UID: "97c071e6-18f7-4a47-9cb1-e74ce7084109"). InnerVolumeSpecName "kube-api-access-dlvcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:15:46 crc kubenswrapper[4744]: I1008 10:15:46.886316 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97c071e6-18f7-4a47-9cb1-e74ce7084109-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97c071e6-18f7-4a47-9cb1-e74ce7084109" (UID: "97c071e6-18f7-4a47-9cb1-e74ce7084109"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:15:46 crc kubenswrapper[4744]: I1008 10:15:46.931350 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97c071e6-18f7-4a47-9cb1-e74ce7084109-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:15:46 crc kubenswrapper[4744]: I1008 10:15:46.931440 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97c071e6-18f7-4a47-9cb1-e74ce7084109-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:15:46 crc kubenswrapper[4744]: I1008 10:15:46.931516 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlvcn\" (UniqueName: \"kubernetes.io/projected/97c071e6-18f7-4a47-9cb1-e74ce7084109-kube-api-access-dlvcn\") on node \"crc\" DevicePath \"\"" Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.202199 4744 generic.go:334] "Generic (PLEG): container finished" podID="97c071e6-18f7-4a47-9cb1-e74ce7084109" containerID="05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6" exitCode=0 Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.202244 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m8smv" event={"ID":"97c071e6-18f7-4a47-9cb1-e74ce7084109","Type":"ContainerDied","Data":"05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6"} Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.202263 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m8smv" Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.202274 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m8smv" event={"ID":"97c071e6-18f7-4a47-9cb1-e74ce7084109","Type":"ContainerDied","Data":"b008a8502b2b6dfaac9fd89fc5aa7541a5e55870460d968febf59833c0979bd7"} Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.202295 4744 scope.go:117] "RemoveContainer" containerID="05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6" Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.226810 4744 scope.go:117] "RemoveContainer" containerID="1052d7d8b79a555fb4282f3560936eb1292d7b95c05cbb4ce2332d2af418f118" Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.244015 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m8smv"] Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.251226 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m8smv"] Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.253578 4744 scope.go:117] "RemoveContainer" containerID="3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56" Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.303977 4744 scope.go:117] "RemoveContainer" containerID="05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6" Oct 08 10:15:47 crc kubenswrapper[4744]: E1008 10:15:47.304709 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6\": container with ID starting with 05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6 not found: ID does not exist" containerID="05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6" Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.304762 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6"} err="failed to get container status \"05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6\": rpc error: code = NotFound desc = could not find container \"05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6\": container with ID starting with 05996c21033bf27378a84d1fd10ff87452b4cab98bcaf8354ab954efbcab34d6 not found: ID does not exist" Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.304794 4744 scope.go:117] "RemoveContainer" containerID="1052d7d8b79a555fb4282f3560936eb1292d7b95c05cbb4ce2332d2af418f118" Oct 08 10:15:47 crc kubenswrapper[4744]: E1008 10:15:47.305094 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1052d7d8b79a555fb4282f3560936eb1292d7b95c05cbb4ce2332d2af418f118\": container with ID starting with 1052d7d8b79a555fb4282f3560936eb1292d7b95c05cbb4ce2332d2af418f118 not found: ID does not exist" containerID="1052d7d8b79a555fb4282f3560936eb1292d7b95c05cbb4ce2332d2af418f118" Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.305123 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1052d7d8b79a555fb4282f3560936eb1292d7b95c05cbb4ce2332d2af418f118"} err="failed to get container status \"1052d7d8b79a555fb4282f3560936eb1292d7b95c05cbb4ce2332d2af418f118\": rpc error: code = NotFound desc = could not find container \"1052d7d8b79a555fb4282f3560936eb1292d7b95c05cbb4ce2332d2af418f118\": container with ID starting with 1052d7d8b79a555fb4282f3560936eb1292d7b95c05cbb4ce2332d2af418f118 not found: ID does not exist" Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.305145 4744 scope.go:117] "RemoveContainer" containerID="3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56" Oct 08 10:15:47 crc kubenswrapper[4744]: E1008 10:15:47.305424 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56\": container with ID starting with 3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56 not found: ID does not exist" containerID="3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56" Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.305443 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56"} err="failed to get container status \"3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56\": rpc error: code = NotFound desc = could not find container \"3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56\": container with ID starting with 3b4a0e38b58900aad07174bf6a71193b206afd07a1c84a8c4dd407f266223d56 not found: ID does not exist" Oct 08 10:15:47 crc kubenswrapper[4744]: I1008 10:15:47.465607 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97c071e6-18f7-4a47-9cb1-e74ce7084109" path="/var/lib/kubelet/pods/97c071e6-18f7-4a47-9cb1-e74ce7084109/volumes" Oct 08 10:15:51 crc kubenswrapper[4744]: I1008 10:15:51.453882 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:15:51 crc kubenswrapper[4744]: E1008 10:15:51.454867 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:15:58 crc kubenswrapper[4744]: I1008 10:15:58.591947 4744 scope.go:117] "RemoveContainer" containerID="46c76ad342f1f057138dd6fac2a75bd62c797e077d5ed791bed275fbce6b4bd5" Oct 08 10:15:58 crc kubenswrapper[4744]: I1008 10:15:58.621667 4744 scope.go:117] "RemoveContainer" containerID="d494b0a7331293cdc351c41896e8f704de2faaf348611501ad220aa61e582341" Oct 08 10:15:58 crc kubenswrapper[4744]: I1008 10:15:58.688662 4744 scope.go:117] "RemoveContainer" containerID="d6dede98da6982a5437e6ceaf32ddbaac0e3cc9e6c3119b070f5829ec0c192b6" Oct 08 10:15:58 crc kubenswrapper[4744]: I1008 10:15:58.814043 4744 scope.go:117] "RemoveContainer" containerID="18a5703374f620b27ffab52b4305c8616c73cc6d0743e3763c1d348b2656b366" Oct 08 10:16:02 crc kubenswrapper[4744]: I1008 10:16:02.453266 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:16:02 crc kubenswrapper[4744]: E1008 10:16:02.454182 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:16:13 crc kubenswrapper[4744]: I1008 10:16:13.454555 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:16:13 crc kubenswrapper[4744]: E1008 10:16:13.455423 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:16:26 crc kubenswrapper[4744]: I1008 10:16:26.453329 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:16:26 crc kubenswrapper[4744]: E1008 10:16:26.454022 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:16:40 crc kubenswrapper[4744]: I1008 10:16:40.453686 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:16:40 crc kubenswrapper[4744]: E1008 10:16:40.454748 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:16:54 crc kubenswrapper[4744]: I1008 10:16:54.453103 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:16:54 crc kubenswrapper[4744]: E1008 10:16:54.455181 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:17:07 crc kubenswrapper[4744]: I1008 10:17:07.454079 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:17:07 crc kubenswrapper[4744]: E1008 10:17:07.455124 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:17:21 crc kubenswrapper[4744]: I1008 10:17:21.454201 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:17:21 crc kubenswrapper[4744]: E1008 10:17:21.455847 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:17:34 crc kubenswrapper[4744]: I1008 10:17:34.453747 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:17:34 crc kubenswrapper[4744]: E1008 10:17:34.455703 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:17:49 crc kubenswrapper[4744]: I1008 10:17:49.453647 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:17:49 crc kubenswrapper[4744]: E1008 10:17:49.454324 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:18:04 crc kubenswrapper[4744]: I1008 10:18:04.453995 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:18:04 crc kubenswrapper[4744]: E1008 10:18:04.456094 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:18:17 crc kubenswrapper[4744]: I1008 10:18:17.454170 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:18:17 crc kubenswrapper[4744]: E1008 10:18:17.455635 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.053982 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d5zfw"] Oct 08 10:18:25 crc kubenswrapper[4744]: E1008 10:18:25.055891 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bac70c8-4f11-4834-9958-90c2aec7787f" containerName="extract-content" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.055969 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bac70c8-4f11-4834-9958-90c2aec7787f" containerName="extract-content" Oct 08 10:18:25 crc kubenswrapper[4744]: E1008 10:18:25.056046 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c071e6-18f7-4a47-9cb1-e74ce7084109" containerName="extract-content" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.056108 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c071e6-18f7-4a47-9cb1-e74ce7084109" containerName="extract-content" Oct 08 10:18:25 crc kubenswrapper[4744]: E1008 10:18:25.056170 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bac70c8-4f11-4834-9958-90c2aec7787f" containerName="registry-server" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.056223 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bac70c8-4f11-4834-9958-90c2aec7787f" containerName="registry-server" Oct 08 10:18:25 crc kubenswrapper[4744]: E1008 10:18:25.056286 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c071e6-18f7-4a47-9cb1-e74ce7084109" containerName="extract-utilities" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.056342 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c071e6-18f7-4a47-9cb1-e74ce7084109" containerName="extract-utilities" Oct 08 10:18:25 crc kubenswrapper[4744]: E1008 10:18:25.056426 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97c071e6-18f7-4a47-9cb1-e74ce7084109" containerName="registry-server" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.056490 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="97c071e6-18f7-4a47-9cb1-e74ce7084109" containerName="registry-server" Oct 08 10:18:25 crc kubenswrapper[4744]: E1008 10:18:25.056560 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bac70c8-4f11-4834-9958-90c2aec7787f" containerName="extract-utilities" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.056624 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bac70c8-4f11-4834-9958-90c2aec7787f" containerName="extract-utilities" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.056851 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="97c071e6-18f7-4a47-9cb1-e74ce7084109" containerName="registry-server" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.056926 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bac70c8-4f11-4834-9958-90c2aec7787f" containerName="registry-server" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.058216 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.082969 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5zfw"] Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.232335 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-catalog-content\") pod \"redhat-marketplace-d5zfw\" (UID: \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\") " pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.232428 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjfpb\" (UniqueName: \"kubernetes.io/projected/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-kube-api-access-zjfpb\") pod \"redhat-marketplace-d5zfw\" (UID: \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\") " pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.232518 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-utilities\") pod \"redhat-marketplace-d5zfw\" (UID: \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\") " pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.334558 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-catalog-content\") pod \"redhat-marketplace-d5zfw\" (UID: \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\") " pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.334618 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjfpb\" (UniqueName: \"kubernetes.io/projected/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-kube-api-access-zjfpb\") pod \"redhat-marketplace-d5zfw\" (UID: \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\") " pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.334681 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-utilities\") pod \"redhat-marketplace-d5zfw\" (UID: \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\") " pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.335164 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-utilities\") pod \"redhat-marketplace-d5zfw\" (UID: \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\") " pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.335392 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-catalog-content\") pod \"redhat-marketplace-d5zfw\" (UID: \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\") " pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.367358 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjfpb\" (UniqueName: \"kubernetes.io/projected/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-kube-api-access-zjfpb\") pod \"redhat-marketplace-d5zfw\" (UID: \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\") " pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:25 crc kubenswrapper[4744]: I1008 10:18:25.413837 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:26 crc kubenswrapper[4744]: I1008 10:18:26.228322 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5zfw"] Oct 08 10:18:26 crc kubenswrapper[4744]: I1008 10:18:26.735407 4744 generic.go:334] "Generic (PLEG): container finished" podID="9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" containerID="8d83054dca4b7553cd600b5bb89f9d48c1a4a02dfcb674f4434d291ebfff341d" exitCode=0 Oct 08 10:18:26 crc kubenswrapper[4744]: I1008 10:18:26.735517 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5zfw" event={"ID":"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1","Type":"ContainerDied","Data":"8d83054dca4b7553cd600b5bb89f9d48c1a4a02dfcb674f4434d291ebfff341d"} Oct 08 10:18:26 crc kubenswrapper[4744]: I1008 10:18:26.735727 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5zfw" event={"ID":"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1","Type":"ContainerStarted","Data":"8c6ecdc8254f3b46c0d47c33b1b79e26b51811ccdc9a4588878caee89dd1a13d"} Oct 08 10:18:29 crc kubenswrapper[4744]: I1008 10:18:29.812610 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5zfw" event={"ID":"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1","Type":"ContainerStarted","Data":"ef46a83ab484fe543bde9f2fdddaa66531211ad76e58d8509bed427d3ae4d90c"} Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.449287 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5zrqr"] Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.451730 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.461074 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5zrqr"] Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.637852 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ww8z\" (UniqueName: \"kubernetes.io/projected/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-kube-api-access-7ww8z\") pod \"redhat-operators-5zrqr\" (UID: \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\") " pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.638002 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-catalog-content\") pod \"redhat-operators-5zrqr\" (UID: \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\") " pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.638036 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-utilities\") pod \"redhat-operators-5zrqr\" (UID: \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\") " pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.740019 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-catalog-content\") pod \"redhat-operators-5zrqr\" (UID: \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\") " pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.740085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-utilities\") pod \"redhat-operators-5zrqr\" (UID: \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\") " pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.740149 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ww8z\" (UniqueName: \"kubernetes.io/projected/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-kube-api-access-7ww8z\") pod \"redhat-operators-5zrqr\" (UID: \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\") " pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.740983 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-catalog-content\") pod \"redhat-operators-5zrqr\" (UID: \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\") " pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.741610 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-utilities\") pod \"redhat-operators-5zrqr\" (UID: \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\") " pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.764165 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ww8z\" (UniqueName: \"kubernetes.io/projected/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-kube-api-access-7ww8z\") pod \"redhat-operators-5zrqr\" (UID: \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\") " pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.770087 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.856677 4744 generic.go:334] "Generic (PLEG): container finished" podID="9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" containerID="ef46a83ab484fe543bde9f2fdddaa66531211ad76e58d8509bed427d3ae4d90c" exitCode=0 Oct 08 10:18:30 crc kubenswrapper[4744]: I1008 10:18:30.856724 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5zfw" event={"ID":"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1","Type":"ContainerDied","Data":"ef46a83ab484fe543bde9f2fdddaa66531211ad76e58d8509bed427d3ae4d90c"} Oct 08 10:18:31 crc kubenswrapper[4744]: I1008 10:18:31.337701 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5zrqr"] Oct 08 10:18:31 crc kubenswrapper[4744]: W1008 10:18:31.355279 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18fbc4ea_c42f_40d9_83cc_7ba2de71af39.slice/crio-4ca2f4a2d80f3ab0570b4d27560e53a5d8d3adb4ac8510ea8202e43ff779082a WatchSource:0}: Error finding container 4ca2f4a2d80f3ab0570b4d27560e53a5d8d3adb4ac8510ea8202e43ff779082a: Status 404 returned error can't find the container with id 4ca2f4a2d80f3ab0570b4d27560e53a5d8d3adb4ac8510ea8202e43ff779082a Oct 08 10:18:31 crc kubenswrapper[4744]: I1008 10:18:31.866251 4744 generic.go:334] "Generic (PLEG): container finished" podID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" containerID="3d370d812c26b9a4f220e454586410761de8fae552f5a9d7deae312fa67fc57d" exitCode=0 Oct 08 10:18:31 crc kubenswrapper[4744]: I1008 10:18:31.866343 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zrqr" event={"ID":"18fbc4ea-c42f-40d9-83cc-7ba2de71af39","Type":"ContainerDied","Data":"3d370d812c26b9a4f220e454586410761de8fae552f5a9d7deae312fa67fc57d"} Oct 08 10:18:31 crc kubenswrapper[4744]: I1008 10:18:31.866440 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zrqr" event={"ID":"18fbc4ea-c42f-40d9-83cc-7ba2de71af39","Type":"ContainerStarted","Data":"4ca2f4a2d80f3ab0570b4d27560e53a5d8d3adb4ac8510ea8202e43ff779082a"} Oct 08 10:18:32 crc kubenswrapper[4744]: I1008 10:18:32.452673 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:18:32 crc kubenswrapper[4744]: E1008 10:18:32.453237 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:18:33 crc kubenswrapper[4744]: I1008 10:18:33.884072 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zrqr" event={"ID":"18fbc4ea-c42f-40d9-83cc-7ba2de71af39","Type":"ContainerStarted","Data":"00d38962971c4518efe2a9b543f15a0ec6c9639c2fe87c29273b29dc287c6fdd"} Oct 08 10:18:33 crc kubenswrapper[4744]: I1008 10:18:33.886791 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5zfw" event={"ID":"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1","Type":"ContainerStarted","Data":"ed39adcf5fe66b24caaa7c1749eb12dbe60f29dd012559503bc67c4b7b2d34fe"} Oct 08 10:18:33 crc kubenswrapper[4744]: I1008 10:18:33.928493 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d5zfw" podStartSLOduration=3.071117713 podStartE2EDuration="8.928470015s" podCreationTimestamp="2025-10-08 10:18:25 +0000 UTC" firstStartedPulling="2025-10-08 10:18:26.737035268 +0000 UTC m=+4001.984680507" lastFinishedPulling="2025-10-08 10:18:32.59438757 +0000 UTC m=+4007.842032809" observedRunningTime="2025-10-08 10:18:33.921703687 +0000 UTC m=+4009.169348936" watchObservedRunningTime="2025-10-08 10:18:33.928470015 +0000 UTC m=+4009.176115254" Oct 08 10:18:35 crc kubenswrapper[4744]: I1008 10:18:35.414497 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:35 crc kubenswrapper[4744]: I1008 10:18:35.414756 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:35 crc kubenswrapper[4744]: I1008 10:18:35.474998 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:40 crc kubenswrapper[4744]: I1008 10:18:40.947752 4744 generic.go:334] "Generic (PLEG): container finished" podID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" containerID="00d38962971c4518efe2a9b543f15a0ec6c9639c2fe87c29273b29dc287c6fdd" exitCode=0 Oct 08 10:18:40 crc kubenswrapper[4744]: I1008 10:18:40.947926 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zrqr" event={"ID":"18fbc4ea-c42f-40d9-83cc-7ba2de71af39","Type":"ContainerDied","Data":"00d38962971c4518efe2a9b543f15a0ec6c9639c2fe87c29273b29dc287c6fdd"} Oct 08 10:18:41 crc kubenswrapper[4744]: I1008 10:18:41.960029 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zrqr" event={"ID":"18fbc4ea-c42f-40d9-83cc-7ba2de71af39","Type":"ContainerStarted","Data":"7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123"} Oct 08 10:18:41 crc kubenswrapper[4744]: I1008 10:18:41.983840 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5zrqr" podStartSLOduration=2.357270913 podStartE2EDuration="11.983818771s" podCreationTimestamp="2025-10-08 10:18:30 +0000 UTC" firstStartedPulling="2025-10-08 10:18:31.868545791 +0000 UTC m=+4007.116191030" lastFinishedPulling="2025-10-08 10:18:41.495093649 +0000 UTC m=+4016.742738888" observedRunningTime="2025-10-08 10:18:41.980832588 +0000 UTC m=+4017.228477857" watchObservedRunningTime="2025-10-08 10:18:41.983818771 +0000 UTC m=+4017.231464010" Oct 08 10:18:45 crc kubenswrapper[4744]: I1008 10:18:45.471798 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:46 crc kubenswrapper[4744]: I1008 10:18:46.628354 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5zfw"] Oct 08 10:18:46 crc kubenswrapper[4744]: I1008 10:18:46.628584 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d5zfw" podUID="9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" containerName="registry-server" containerID="cri-o://ed39adcf5fe66b24caaa7c1749eb12dbe60f29dd012559503bc67c4b7b2d34fe" gracePeriod=2 Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.006086 4744 generic.go:334] "Generic (PLEG): container finished" podID="9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" containerID="ed39adcf5fe66b24caaa7c1749eb12dbe60f29dd012559503bc67c4b7b2d34fe" exitCode=0 Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.006423 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5zfw" event={"ID":"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1","Type":"ContainerDied","Data":"ed39adcf5fe66b24caaa7c1749eb12dbe60f29dd012559503bc67c4b7b2d34fe"} Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.175593 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.297437 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjfpb\" (UniqueName: \"kubernetes.io/projected/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-kube-api-access-zjfpb\") pod \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\" (UID: \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\") " Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.297566 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-utilities\") pod \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\" (UID: \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\") " Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.297699 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-catalog-content\") pod \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\" (UID: \"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1\") " Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.301097 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-utilities" (OuterVolumeSpecName: "utilities") pod "9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" (UID: "9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.308886 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-kube-api-access-zjfpb" (OuterVolumeSpecName: "kube-api-access-zjfpb") pod "9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" (UID: "9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1"). InnerVolumeSpecName "kube-api-access-zjfpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.312309 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" (UID: "9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.399772 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.399816 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjfpb\" (UniqueName: \"kubernetes.io/projected/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-kube-api-access-zjfpb\") on node \"crc\" DevicePath \"\"" Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.399833 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:18:47 crc kubenswrapper[4744]: I1008 10:18:47.453629 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:18:47 crc kubenswrapper[4744]: E1008 10:18:47.453874 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:18:48 crc kubenswrapper[4744]: I1008 10:18:48.023403 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5zfw" event={"ID":"9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1","Type":"ContainerDied","Data":"8c6ecdc8254f3b46c0d47c33b1b79e26b51811ccdc9a4588878caee89dd1a13d"} Oct 08 10:18:48 crc kubenswrapper[4744]: I1008 10:18:48.023462 4744 scope.go:117] "RemoveContainer" containerID="ed39adcf5fe66b24caaa7c1749eb12dbe60f29dd012559503bc67c4b7b2d34fe" Oct 08 10:18:48 crc kubenswrapper[4744]: I1008 10:18:48.024219 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d5zfw" Oct 08 10:18:48 crc kubenswrapper[4744]: I1008 10:18:48.051313 4744 scope.go:117] "RemoveContainer" containerID="ef46a83ab484fe543bde9f2fdddaa66531211ad76e58d8509bed427d3ae4d90c" Oct 08 10:18:48 crc kubenswrapper[4744]: I1008 10:18:48.055328 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5zfw"] Oct 08 10:18:48 crc kubenswrapper[4744]: I1008 10:18:48.064780 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5zfw"] Oct 08 10:18:48 crc kubenswrapper[4744]: I1008 10:18:48.076966 4744 scope.go:117] "RemoveContainer" containerID="8d83054dca4b7553cd600b5bb89f9d48c1a4a02dfcb674f4434d291ebfff341d" Oct 08 10:18:49 crc kubenswrapper[4744]: I1008 10:18:49.465065 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" path="/var/lib/kubelet/pods/9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1/volumes" Oct 08 10:18:50 crc kubenswrapper[4744]: I1008 10:18:50.770221 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:50 crc kubenswrapper[4744]: I1008 10:18:50.770630 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:18:51 crc kubenswrapper[4744]: I1008 10:18:51.820030 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5zrqr" podUID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" containerName="registry-server" probeResult="failure" output=< Oct 08 10:18:51 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 10:18:51 crc kubenswrapper[4744]: > Oct 08 10:18:58 crc kubenswrapper[4744]: I1008 10:18:58.454521 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:18:59 crc kubenswrapper[4744]: I1008 10:18:59.152478 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"ca8fafcf7cc39e7299b68be5cb7e1e887b65f71741164b2f539597d73f629078"} Oct 08 10:19:00 crc kubenswrapper[4744]: I1008 10:19:00.831248 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:19:00 crc kubenswrapper[4744]: I1008 10:19:00.898623 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:19:01 crc kubenswrapper[4744]: I1008 10:19:01.642999 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5zrqr"] Oct 08 10:19:02 crc kubenswrapper[4744]: I1008 10:19:02.184401 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5zrqr" podUID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" containerName="registry-server" containerID="cri-o://7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123" gracePeriod=2 Oct 08 10:19:02 crc kubenswrapper[4744]: I1008 10:19:02.768115 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:19:02 crc kubenswrapper[4744]: I1008 10:19:02.926185 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-catalog-content\") pod \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\" (UID: \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\") " Oct 08 10:19:02 crc kubenswrapper[4744]: I1008 10:19:02.931582 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-utilities\") pod \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\" (UID: \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\") " Oct 08 10:19:02 crc kubenswrapper[4744]: I1008 10:19:02.931764 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ww8z\" (UniqueName: \"kubernetes.io/projected/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-kube-api-access-7ww8z\") pod \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\" (UID: \"18fbc4ea-c42f-40d9-83cc-7ba2de71af39\") " Oct 08 10:19:02 crc kubenswrapper[4744]: I1008 10:19:02.932661 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-utilities" (OuterVolumeSpecName: "utilities") pod "18fbc4ea-c42f-40d9-83cc-7ba2de71af39" (UID: "18fbc4ea-c42f-40d9-83cc-7ba2de71af39"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:19:02 crc kubenswrapper[4744]: I1008 10:19:02.933067 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:19:02 crc kubenswrapper[4744]: I1008 10:19:02.938872 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-kube-api-access-7ww8z" (OuterVolumeSpecName: "kube-api-access-7ww8z") pod "18fbc4ea-c42f-40d9-83cc-7ba2de71af39" (UID: "18fbc4ea-c42f-40d9-83cc-7ba2de71af39"). InnerVolumeSpecName "kube-api-access-7ww8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.036172 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ww8z\" (UniqueName: \"kubernetes.io/projected/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-kube-api-access-7ww8z\") on node \"crc\" DevicePath \"\"" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.038339 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18fbc4ea-c42f-40d9-83cc-7ba2de71af39" (UID: "18fbc4ea-c42f-40d9-83cc-7ba2de71af39"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.138055 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18fbc4ea-c42f-40d9-83cc-7ba2de71af39-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.207030 4744 generic.go:334] "Generic (PLEG): container finished" podID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" containerID="7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123" exitCode=0 Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.207083 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zrqr" event={"ID":"18fbc4ea-c42f-40d9-83cc-7ba2de71af39","Type":"ContainerDied","Data":"7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123"} Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.207116 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zrqr" event={"ID":"18fbc4ea-c42f-40d9-83cc-7ba2de71af39","Type":"ContainerDied","Data":"4ca2f4a2d80f3ab0570b4d27560e53a5d8d3adb4ac8510ea8202e43ff779082a"} Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.207135 4744 scope.go:117] "RemoveContainer" containerID="7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.207091 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zrqr" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.247602 4744 scope.go:117] "RemoveContainer" containerID="00d38962971c4518efe2a9b543f15a0ec6c9639c2fe87c29273b29dc287c6fdd" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.253443 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5zrqr"] Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.267202 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5zrqr"] Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.271865 4744 scope.go:117] "RemoveContainer" containerID="3d370d812c26b9a4f220e454586410761de8fae552f5a9d7deae312fa67fc57d" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.311355 4744 scope.go:117] "RemoveContainer" containerID="7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123" Oct 08 10:19:03 crc kubenswrapper[4744]: E1008 10:19:03.312118 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123\": container with ID starting with 7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123 not found: ID does not exist" containerID="7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.312157 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123"} err="failed to get container status \"7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123\": rpc error: code = NotFound desc = could not find container \"7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123\": container with ID starting with 7b43a133d3d12b5972c5d6071ebbe262eb342c4b1cc66380f527c777dc138123 not found: ID does not exist" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.312180 4744 scope.go:117] "RemoveContainer" containerID="00d38962971c4518efe2a9b543f15a0ec6c9639c2fe87c29273b29dc287c6fdd" Oct 08 10:19:03 crc kubenswrapper[4744]: E1008 10:19:03.312651 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00d38962971c4518efe2a9b543f15a0ec6c9639c2fe87c29273b29dc287c6fdd\": container with ID starting with 00d38962971c4518efe2a9b543f15a0ec6c9639c2fe87c29273b29dc287c6fdd not found: ID does not exist" containerID="00d38962971c4518efe2a9b543f15a0ec6c9639c2fe87c29273b29dc287c6fdd" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.312708 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00d38962971c4518efe2a9b543f15a0ec6c9639c2fe87c29273b29dc287c6fdd"} err="failed to get container status \"00d38962971c4518efe2a9b543f15a0ec6c9639c2fe87c29273b29dc287c6fdd\": rpc error: code = NotFound desc = could not find container \"00d38962971c4518efe2a9b543f15a0ec6c9639c2fe87c29273b29dc287c6fdd\": container with ID starting with 00d38962971c4518efe2a9b543f15a0ec6c9639c2fe87c29273b29dc287c6fdd not found: ID does not exist" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.312743 4744 scope.go:117] "RemoveContainer" containerID="3d370d812c26b9a4f220e454586410761de8fae552f5a9d7deae312fa67fc57d" Oct 08 10:19:03 crc kubenswrapper[4744]: E1008 10:19:03.313168 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d370d812c26b9a4f220e454586410761de8fae552f5a9d7deae312fa67fc57d\": container with ID starting with 3d370d812c26b9a4f220e454586410761de8fae552f5a9d7deae312fa67fc57d not found: ID does not exist" containerID="3d370d812c26b9a4f220e454586410761de8fae552f5a9d7deae312fa67fc57d" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.313235 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d370d812c26b9a4f220e454586410761de8fae552f5a9d7deae312fa67fc57d"} err="failed to get container status \"3d370d812c26b9a4f220e454586410761de8fae552f5a9d7deae312fa67fc57d\": rpc error: code = NotFound desc = could not find container \"3d370d812c26b9a4f220e454586410761de8fae552f5a9d7deae312fa67fc57d\": container with ID starting with 3d370d812c26b9a4f220e454586410761de8fae552f5a9d7deae312fa67fc57d not found: ID does not exist" Oct 08 10:19:03 crc kubenswrapper[4744]: I1008 10:19:03.464704 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" path="/var/lib/kubelet/pods/18fbc4ea-c42f-40d9-83cc-7ba2de71af39/volumes" Oct 08 10:21:19 crc kubenswrapper[4744]: I1008 10:21:19.690859 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:21:19 crc kubenswrapper[4744]: I1008 10:21:19.691346 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:21:36 crc kubenswrapper[4744]: I1008 10:21:36.641445 4744 generic.go:334] "Generic (PLEG): container finished" podID="ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" containerID="b7c39adb1c4c22a29d29fad2de16bf87f1c7cef29adf3e343f39ee82be31941f" exitCode=0 Oct 08 10:21:36 crc kubenswrapper[4744]: I1008 10:21:36.641514 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be","Type":"ContainerDied","Data":"b7c39adb1c4c22a29d29fad2de16bf87f1c7cef29adf3e343f39ee82be31941f"} Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.093814 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.224805 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-openstack-config-secret\") pod \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.224901 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h776v\" (UniqueName: \"kubernetes.io/projected/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-kube-api-access-h776v\") pod \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.224926 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-config-data\") pod \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.225285 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-ca-certs\") pod \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.225356 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-ssh-key\") pod \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.225576 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.225634 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-test-operator-ephemeral-temporary\") pod \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.225700 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-openstack-config\") pod \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.225809 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-test-operator-ephemeral-workdir\") pod \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\" (UID: \"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be\") " Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.226202 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-config-data" (OuterVolumeSpecName: "config-data") pod "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" (UID: "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.227091 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" (UID: "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.227127 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.233570 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" (UID: "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.234078 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-kube-api-access-h776v" (OuterVolumeSpecName: "kube-api-access-h776v") pod "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" (UID: "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be"). InnerVolumeSpecName "kube-api-access-h776v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.236798 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" (UID: "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.259008 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" (UID: "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.265322 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" (UID: "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.266328 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" (UID: "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.284597 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" (UID: "ba86d64c-cc8a-48b2-b70e-2ca6cf2145be"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.330169 4744 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.330215 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.330231 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h776v\" (UniqueName: \"kubernetes.io/projected/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-kube-api-access-h776v\") on node \"crc\" DevicePath \"\"" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.330245 4744 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.330260 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.330867 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.330898 4744 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.330915 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ba86d64c-cc8a-48b2-b70e-2ca6cf2145be-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.360828 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.433408 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.667907 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ba86d64c-cc8a-48b2-b70e-2ca6cf2145be","Type":"ContainerDied","Data":"fd623e980c7ee1817d3a8eb3ad7c3aa7e53ed23408f0966446fbe864a4afea02"} Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.668302 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd623e980c7ee1817d3a8eb3ad7c3aa7e53ed23408f0966446fbe864a4afea02" Oct 08 10:21:38 crc kubenswrapper[4744]: I1008 10:21:38.668369 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.079133 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 08 10:21:44 crc kubenswrapper[4744]: E1008 10:21:44.080296 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" containerName="extract-content" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.080310 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" containerName="extract-content" Oct 08 10:21:44 crc kubenswrapper[4744]: E1008 10:21:44.080325 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" containerName="registry-server" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.080331 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" containerName="registry-server" Oct 08 10:21:44 crc kubenswrapper[4744]: E1008 10:21:44.080342 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" containerName="tempest-tests-tempest-tests-runner" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.080350 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" containerName="tempest-tests-tempest-tests-runner" Oct 08 10:21:44 crc kubenswrapper[4744]: E1008 10:21:44.080447 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" containerName="registry-server" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.080455 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" containerName="registry-server" Oct 08 10:21:44 crc kubenswrapper[4744]: E1008 10:21:44.080475 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" containerName="extract-utilities" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.080482 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" containerName="extract-utilities" Oct 08 10:21:44 crc kubenswrapper[4744]: E1008 10:21:44.080497 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" containerName="extract-content" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.080503 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" containerName="extract-content" Oct 08 10:21:44 crc kubenswrapper[4744]: E1008 10:21:44.080511 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" containerName="extract-utilities" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.080517 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" containerName="extract-utilities" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.080709 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba86d64c-cc8a-48b2-b70e-2ca6cf2145be" containerName="tempest-tests-tempest-tests-runner" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.080731 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bc0aed7-7443-43d6-a0ab-9fe0f99c6aa1" containerName="registry-server" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.080740 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="18fbc4ea-c42f-40d9-83cc-7ba2de71af39" containerName="registry-server" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.081424 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.088822 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-g5r6h" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.097174 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.169923 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2ng9\" (UniqueName: \"kubernetes.io/projected/e36a0230-d350-400e-a974-ce78f036dd64-kube-api-access-v2ng9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e36a0230-d350-400e-a974-ce78f036dd64\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.170105 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e36a0230-d350-400e-a974-ce78f036dd64\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.272966 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2ng9\" (UniqueName: \"kubernetes.io/projected/e36a0230-d350-400e-a974-ce78f036dd64-kube-api-access-v2ng9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e36a0230-d350-400e-a974-ce78f036dd64\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.273070 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e36a0230-d350-400e-a974-ce78f036dd64\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.273967 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e36a0230-d350-400e-a974-ce78f036dd64\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.314878 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2ng9\" (UniqueName: \"kubernetes.io/projected/e36a0230-d350-400e-a974-ce78f036dd64-kube-api-access-v2ng9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e36a0230-d350-400e-a974-ce78f036dd64\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.398078 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e36a0230-d350-400e-a974-ce78f036dd64\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.407060 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.896255 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 08 10:21:44 crc kubenswrapper[4744]: I1008 10:21:44.923712 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 10:21:45 crc kubenswrapper[4744]: I1008 10:21:45.741994 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"e36a0230-d350-400e-a974-ce78f036dd64","Type":"ContainerStarted","Data":"d58e31cb40f5cd5e87e9d34864912d0d5fdaebd3d8e63e9a0550cee2463a9e0e"} Oct 08 10:21:47 crc kubenswrapper[4744]: I1008 10:21:47.758675 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"e36a0230-d350-400e-a974-ce78f036dd64","Type":"ContainerStarted","Data":"46e8c80acea0e087ff4b398aeba200187f6799c1d09c8e7607067f255a5209cd"} Oct 08 10:21:47 crc kubenswrapper[4744]: I1008 10:21:47.774454 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.745399528 podStartE2EDuration="3.774434325s" podCreationTimestamp="2025-10-08 10:21:44 +0000 UTC" firstStartedPulling="2025-10-08 10:21:44.923023361 +0000 UTC m=+4200.170668600" lastFinishedPulling="2025-10-08 10:21:46.952058158 +0000 UTC m=+4202.199703397" observedRunningTime="2025-10-08 10:21:47.773780897 +0000 UTC m=+4203.021426136" watchObservedRunningTime="2025-10-08 10:21:47.774434325 +0000 UTC m=+4203.022079564" Oct 08 10:21:49 crc kubenswrapper[4744]: I1008 10:21:49.690355 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:21:49 crc kubenswrapper[4744]: I1008 10:21:49.690701 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:22:05 crc kubenswrapper[4744]: I1008 10:22:05.542002 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lx88m/must-gather-wrnk6"] Oct 08 10:22:05 crc kubenswrapper[4744]: I1008 10:22:05.543913 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/must-gather-wrnk6" Oct 08 10:22:05 crc kubenswrapper[4744]: I1008 10:22:05.550782 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lx88m"/"kube-root-ca.crt" Oct 08 10:22:05 crc kubenswrapper[4744]: I1008 10:22:05.551282 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lx88m"/"openshift-service-ca.crt" Oct 08 10:22:05 crc kubenswrapper[4744]: I1008 10:22:05.564745 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lx88m/must-gather-wrnk6"] Oct 08 10:22:05 crc kubenswrapper[4744]: I1008 10:22:05.661807 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a365de61-417b-4735-b90a-e8a8feb93085-must-gather-output\") pod \"must-gather-wrnk6\" (UID: \"a365de61-417b-4735-b90a-e8a8feb93085\") " pod="openshift-must-gather-lx88m/must-gather-wrnk6" Oct 08 10:22:05 crc kubenswrapper[4744]: I1008 10:22:05.661924 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9vn4\" (UniqueName: \"kubernetes.io/projected/a365de61-417b-4735-b90a-e8a8feb93085-kube-api-access-r9vn4\") pod \"must-gather-wrnk6\" (UID: \"a365de61-417b-4735-b90a-e8a8feb93085\") " pod="openshift-must-gather-lx88m/must-gather-wrnk6" Oct 08 10:22:05 crc kubenswrapper[4744]: I1008 10:22:05.763642 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9vn4\" (UniqueName: \"kubernetes.io/projected/a365de61-417b-4735-b90a-e8a8feb93085-kube-api-access-r9vn4\") pod \"must-gather-wrnk6\" (UID: \"a365de61-417b-4735-b90a-e8a8feb93085\") " pod="openshift-must-gather-lx88m/must-gather-wrnk6" Oct 08 10:22:05 crc kubenswrapper[4744]: I1008 10:22:05.763755 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a365de61-417b-4735-b90a-e8a8feb93085-must-gather-output\") pod \"must-gather-wrnk6\" (UID: \"a365de61-417b-4735-b90a-e8a8feb93085\") " pod="openshift-must-gather-lx88m/must-gather-wrnk6" Oct 08 10:22:05 crc kubenswrapper[4744]: I1008 10:22:05.764161 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a365de61-417b-4735-b90a-e8a8feb93085-must-gather-output\") pod \"must-gather-wrnk6\" (UID: \"a365de61-417b-4735-b90a-e8a8feb93085\") " pod="openshift-must-gather-lx88m/must-gather-wrnk6" Oct 08 10:22:05 crc kubenswrapper[4744]: I1008 10:22:05.788021 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9vn4\" (UniqueName: \"kubernetes.io/projected/a365de61-417b-4735-b90a-e8a8feb93085-kube-api-access-r9vn4\") pod \"must-gather-wrnk6\" (UID: \"a365de61-417b-4735-b90a-e8a8feb93085\") " pod="openshift-must-gather-lx88m/must-gather-wrnk6" Oct 08 10:22:05 crc kubenswrapper[4744]: I1008 10:22:05.878211 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/must-gather-wrnk6" Oct 08 10:22:06 crc kubenswrapper[4744]: I1008 10:22:06.348704 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lx88m/must-gather-wrnk6"] Oct 08 10:22:06 crc kubenswrapper[4744]: W1008 10:22:06.349521 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda365de61_417b_4735_b90a_e8a8feb93085.slice/crio-de62870250023a3ce96aecd2b48d306d280a2365af0418ac5672e79469deee03 WatchSource:0}: Error finding container de62870250023a3ce96aecd2b48d306d280a2365af0418ac5672e79469deee03: Status 404 returned error can't find the container with id de62870250023a3ce96aecd2b48d306d280a2365af0418ac5672e79469deee03 Oct 08 10:22:06 crc kubenswrapper[4744]: I1008 10:22:06.942155 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/must-gather-wrnk6" event={"ID":"a365de61-417b-4735-b90a-e8a8feb93085","Type":"ContainerStarted","Data":"de62870250023a3ce96aecd2b48d306d280a2365af0418ac5672e79469deee03"} Oct 08 10:22:12 crc kubenswrapper[4744]: I1008 10:22:12.000537 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/must-gather-wrnk6" event={"ID":"a365de61-417b-4735-b90a-e8a8feb93085","Type":"ContainerStarted","Data":"a19df090fa08ea0ecad3c06c82d8f110bc82ab2e7b14404d061e3b2acab79a9f"} Oct 08 10:22:12 crc kubenswrapper[4744]: I1008 10:22:12.001101 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/must-gather-wrnk6" event={"ID":"a365de61-417b-4735-b90a-e8a8feb93085","Type":"ContainerStarted","Data":"f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1"} Oct 08 10:22:12 crc kubenswrapper[4744]: I1008 10:22:12.015206 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lx88m/must-gather-wrnk6" podStartSLOduration=2.101258878 podStartE2EDuration="7.015190718s" podCreationTimestamp="2025-10-08 10:22:05 +0000 UTC" firstStartedPulling="2025-10-08 10:22:06.351102267 +0000 UTC m=+4221.598747516" lastFinishedPulling="2025-10-08 10:22:11.265034077 +0000 UTC m=+4226.512679356" observedRunningTime="2025-10-08 10:22:12.013136131 +0000 UTC m=+4227.260781370" watchObservedRunningTime="2025-10-08 10:22:12.015190718 +0000 UTC m=+4227.262835957" Oct 08 10:22:18 crc kubenswrapper[4744]: I1008 10:22:18.227114 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lx88m/crc-debug-2794j"] Oct 08 10:22:18 crc kubenswrapper[4744]: I1008 10:22:18.229206 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/crc-debug-2794j" Oct 08 10:22:18 crc kubenswrapper[4744]: I1008 10:22:18.232693 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-lx88m"/"default-dockercfg-mmpt4" Oct 08 10:22:18 crc kubenswrapper[4744]: I1008 10:22:18.319810 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c892292-d161-4d31-b101-4f2af00364a1-host\") pod \"crc-debug-2794j\" (UID: \"9c892292-d161-4d31-b101-4f2af00364a1\") " pod="openshift-must-gather-lx88m/crc-debug-2794j" Oct 08 10:22:18 crc kubenswrapper[4744]: I1008 10:22:18.319900 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m97hr\" (UniqueName: \"kubernetes.io/projected/9c892292-d161-4d31-b101-4f2af00364a1-kube-api-access-m97hr\") pod \"crc-debug-2794j\" (UID: \"9c892292-d161-4d31-b101-4f2af00364a1\") " pod="openshift-must-gather-lx88m/crc-debug-2794j" Oct 08 10:22:18 crc kubenswrapper[4744]: I1008 10:22:18.422155 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c892292-d161-4d31-b101-4f2af00364a1-host\") pod \"crc-debug-2794j\" (UID: \"9c892292-d161-4d31-b101-4f2af00364a1\") " pod="openshift-must-gather-lx88m/crc-debug-2794j" Oct 08 10:22:18 crc kubenswrapper[4744]: I1008 10:22:18.422196 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m97hr\" (UniqueName: \"kubernetes.io/projected/9c892292-d161-4d31-b101-4f2af00364a1-kube-api-access-m97hr\") pod \"crc-debug-2794j\" (UID: \"9c892292-d161-4d31-b101-4f2af00364a1\") " pod="openshift-must-gather-lx88m/crc-debug-2794j" Oct 08 10:22:18 crc kubenswrapper[4744]: I1008 10:22:18.422319 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c892292-d161-4d31-b101-4f2af00364a1-host\") pod \"crc-debug-2794j\" (UID: \"9c892292-d161-4d31-b101-4f2af00364a1\") " pod="openshift-must-gather-lx88m/crc-debug-2794j" Oct 08 10:22:18 crc kubenswrapper[4744]: I1008 10:22:18.448179 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m97hr\" (UniqueName: \"kubernetes.io/projected/9c892292-d161-4d31-b101-4f2af00364a1-kube-api-access-m97hr\") pod \"crc-debug-2794j\" (UID: \"9c892292-d161-4d31-b101-4f2af00364a1\") " pod="openshift-must-gather-lx88m/crc-debug-2794j" Oct 08 10:22:18 crc kubenswrapper[4744]: I1008 10:22:18.549496 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/crc-debug-2794j" Oct 08 10:22:19 crc kubenswrapper[4744]: I1008 10:22:19.067049 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/crc-debug-2794j" event={"ID":"9c892292-d161-4d31-b101-4f2af00364a1","Type":"ContainerStarted","Data":"9aa69b358fe4e9cd643b64103fdcf3cf36bbc308cad34123bfc6aac85b55ce02"} Oct 08 10:22:19 crc kubenswrapper[4744]: I1008 10:22:19.690226 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:22:19 crc kubenswrapper[4744]: I1008 10:22:19.690298 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:22:19 crc kubenswrapper[4744]: I1008 10:22:19.690345 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 10:22:19 crc kubenswrapper[4744]: I1008 10:22:19.691160 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca8fafcf7cc39e7299b68be5cb7e1e887b65f71741164b2f539597d73f629078"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 10:22:19 crc kubenswrapper[4744]: I1008 10:22:19.691216 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://ca8fafcf7cc39e7299b68be5cb7e1e887b65f71741164b2f539597d73f629078" gracePeriod=600 Oct 08 10:22:20 crc kubenswrapper[4744]: I1008 10:22:20.080290 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="ca8fafcf7cc39e7299b68be5cb7e1e887b65f71741164b2f539597d73f629078" exitCode=0 Oct 08 10:22:20 crc kubenswrapper[4744]: I1008 10:22:20.080494 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"ca8fafcf7cc39e7299b68be5cb7e1e887b65f71741164b2f539597d73f629078"} Oct 08 10:22:20 crc kubenswrapper[4744]: I1008 10:22:20.080627 4744 scope.go:117] "RemoveContainer" containerID="0bbfff0cc056019951b644d34d8385d98f916430fb2e823842513c652372390a" Oct 08 10:22:21 crc kubenswrapper[4744]: I1008 10:22:21.092510 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab"} Oct 08 10:22:32 crc kubenswrapper[4744]: I1008 10:22:32.193551 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/crc-debug-2794j" event={"ID":"9c892292-d161-4d31-b101-4f2af00364a1","Type":"ContainerStarted","Data":"2ce2e5bed34ee83a1fba64376cde9515a540d0409a38130a83dde8ee3d95eb0c"} Oct 08 10:22:32 crc kubenswrapper[4744]: I1008 10:22:32.212707 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lx88m/crc-debug-2794j" podStartSLOduration=1.5606504829999999 podStartE2EDuration="14.21268781s" podCreationTimestamp="2025-10-08 10:22:18 +0000 UTC" firstStartedPulling="2025-10-08 10:22:18.58390505 +0000 UTC m=+4233.831550289" lastFinishedPulling="2025-10-08 10:22:31.235942377 +0000 UTC m=+4246.483587616" observedRunningTime="2025-10-08 10:22:32.208009041 +0000 UTC m=+4247.455654290" watchObservedRunningTime="2025-10-08 10:22:32.21268781 +0000 UTC m=+4247.460333039" Oct 08 10:24:07 crc kubenswrapper[4744]: I1008 10:24:07.911414 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6d98c4654d-2xrfx_ea9a2551-0abd-4bce-9785-4c8c7e11b02d/barbican-api/0.log" Oct 08 10:24:07 crc kubenswrapper[4744]: I1008 10:24:07.952094 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6d98c4654d-2xrfx_ea9a2551-0abd-4bce-9785-4c8c7e11b02d/barbican-api-log/0.log" Oct 08 10:24:08 crc kubenswrapper[4744]: I1008 10:24:08.278679 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-8578895ff6-xg92r_4ca73575-280f-43af-8c34-123a1cf83e92/barbican-keystone-listener/0.log" Oct 08 10:24:08 crc kubenswrapper[4744]: I1008 10:24:08.368942 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-8578895ff6-xg92r_4ca73575-280f-43af-8c34-123a1cf83e92/barbican-keystone-listener-log/0.log" Oct 08 10:24:08 crc kubenswrapper[4744]: I1008 10:24:08.518057 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5ccc6f877f-j2zc9_203b8276-cafa-4b89-a68d-0f9437ee7161/barbican-worker/0.log" Oct 08 10:24:08 crc kubenswrapper[4744]: I1008 10:24:08.568954 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5ccc6f877f-j2zc9_203b8276-cafa-4b89-a68d-0f9437ee7161/barbican-worker-log/0.log" Oct 08 10:24:09 crc kubenswrapper[4744]: I1008 10:24:09.275598 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4_f0464e55-2b52-4c5a-bc87-9c2c87cb6101/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:09 crc kubenswrapper[4744]: I1008 10:24:09.514932 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2568ab3e-d5f6-4524-b1dc-20a6f389cd31/ceilometer-notification-agent/0.log" Oct 08 10:24:09 crc kubenswrapper[4744]: I1008 10:24:09.560877 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2568ab3e-d5f6-4524-b1dc-20a6f389cd31/ceilometer-central-agent/0.log" Oct 08 10:24:09 crc kubenswrapper[4744]: I1008 10:24:09.577400 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2568ab3e-d5f6-4524-b1dc-20a6f389cd31/proxy-httpd/0.log" Oct 08 10:24:09 crc kubenswrapper[4744]: I1008 10:24:09.799436 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2568ab3e-d5f6-4524-b1dc-20a6f389cd31/sg-core/0.log" Oct 08 10:24:09 crc kubenswrapper[4744]: I1008 10:24:09.871350 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k_664b4c04-3a69-444f-aa77-621229ee5b96/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:10 crc kubenswrapper[4744]: I1008 10:24:10.090839 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl_dd384bed-7c22-4e1a-8342-94589991c934/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:10 crc kubenswrapper[4744]: I1008 10:24:10.410520 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9efda746-3b68-4707-bbe8-3f4e74db5a60/cinder-api/0.log" Oct 08 10:24:10 crc kubenswrapper[4744]: I1008 10:24:10.452621 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9efda746-3b68-4707-bbe8-3f4e74db5a60/cinder-api-log/0.log" Oct 08 10:24:10 crc kubenswrapper[4744]: I1008 10:24:10.761887 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_d4fd8637-8969-4c7f-9ecf-9276b5550484/cinder-backup/0.log" Oct 08 10:24:10 crc kubenswrapper[4744]: I1008 10:24:10.779242 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_d4fd8637-8969-4c7f-9ecf-9276b5550484/probe/0.log" Oct 08 10:24:11 crc kubenswrapper[4744]: I1008 10:24:11.257682 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_49177d36-177b-4b09-a93e-3d602ff30a27/cinder-scheduler/0.log" Oct 08 10:24:11 crc kubenswrapper[4744]: I1008 10:24:11.421908 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_49177d36-177b-4b09-a93e-3d602ff30a27/probe/0.log" Oct 08 10:24:11 crc kubenswrapper[4744]: I1008 10:24:11.566026 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_9cc1c714-af01-4205-8670-9055cdb3c623/cinder-volume/0.log" Oct 08 10:24:11 crc kubenswrapper[4744]: I1008 10:24:11.713105 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_9cc1c714-af01-4205-8670-9055cdb3c623/probe/0.log" Oct 08 10:24:12 crc kubenswrapper[4744]: I1008 10:24:12.029055 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-k48qv_d0fc8259-e8cd-4d69-8a95-f74d908d569e/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:12 crc kubenswrapper[4744]: I1008 10:24:12.275666 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-wcf57_bf19b11f-c43e-4f3b-a34c-53d3837705fb/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:12 crc kubenswrapper[4744]: I1008 10:24:12.308722 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7887c4559f-jqcwg_084ffd95-c3f3-4340-bc26-d3c2c427c8ad/init/0.log" Oct 08 10:24:12 crc kubenswrapper[4744]: I1008 10:24:12.755895 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7887c4559f-jqcwg_084ffd95-c3f3-4340-bc26-d3c2c427c8ad/init/0.log" Oct 08 10:24:12 crc kubenswrapper[4744]: I1008 10:24:12.818662 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7887c4559f-jqcwg_084ffd95-c3f3-4340-bc26-d3c2c427c8ad/dnsmasq-dns/0.log" Oct 08 10:24:12 crc kubenswrapper[4744]: I1008 10:24:12.843921 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_26c9e730-4ef5-41f3-b861-fbe778351306/glance-httpd/0.log" Oct 08 10:24:13 crc kubenswrapper[4744]: I1008 10:24:13.047897 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_26c9e730-4ef5-41f3-b861-fbe778351306/glance-log/0.log" Oct 08 10:24:13 crc kubenswrapper[4744]: I1008 10:24:13.214355 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2bdd463c-e43b-4740-ba42-3981cb6b1380/glance-httpd/0.log" Oct 08 10:24:13 crc kubenswrapper[4744]: I1008 10:24:13.327213 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2bdd463c-e43b-4740-ba42-3981cb6b1380/glance-log/0.log" Oct 08 10:24:13 crc kubenswrapper[4744]: I1008 10:24:13.736363 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5549798486-rbbgm_f337fa5f-1088-407d-a832-856ff504cd72/horizon/0.log" Oct 08 10:24:13 crc kubenswrapper[4744]: I1008 10:24:13.903956 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5549798486-rbbgm_f337fa5f-1088-407d-a832-856ff504cd72/horizon-log/0.log" Oct 08 10:24:14 crc kubenswrapper[4744]: I1008 10:24:14.186607 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bctfd_3876b90a-308b-4320-a8ab-f81a48295c1f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:14 crc kubenswrapper[4744]: I1008 10:24:14.297331 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-cd6ks_57e3beb8-aeaa-435a-aa5a-78f854db1d6d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:14 crc kubenswrapper[4744]: I1008 10:24:14.566675 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29331961-k29hs_6d3e0b57-177d-4351-8fce-3a758311bb35/keystone-cron/0.log" Oct 08 10:24:14 crc kubenswrapper[4744]: I1008 10:24:14.667236 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c4d7b4588-6wcr4_2af8995b-abf6-4615-bfda-f5900a1bfb2c/keystone-api/0.log" Oct 08 10:24:14 crc kubenswrapper[4744]: I1008 10:24:14.853183 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a8b493e3-c187-4150-95db-3a4f110d15ea/kube-state-metrics/0.log" Oct 08 10:24:14 crc kubenswrapper[4744]: I1008 10:24:14.963973 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-swsrc_e80d5289-7657-4835-8b43-936537129935/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:15 crc kubenswrapper[4744]: I1008 10:24:15.244152 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_1003e3fc-33a3-45d1-a1b7-9f383af6e336/manila-api/0.log" Oct 08 10:24:15 crc kubenswrapper[4744]: I1008 10:24:15.282044 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_1003e3fc-33a3-45d1-a1b7-9f383af6e336/manila-api-log/0.log" Oct 08 10:24:15 crc kubenswrapper[4744]: I1008 10:24:15.513423 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e/manila-scheduler/0.log" Oct 08 10:24:15 crc kubenswrapper[4744]: I1008 10:24:15.526301 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e/probe/0.log" Oct 08 10:24:15 crc kubenswrapper[4744]: I1008 10:24:15.689848 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_0e308a1c-d62d-4782-adba-7517aab54f94/manila-share/0.log" Oct 08 10:24:15 crc kubenswrapper[4744]: I1008 10:24:15.746885 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_0e308a1c-d62d-4782-adba-7517aab54f94/probe/0.log" Oct 08 10:24:16 crc kubenswrapper[4744]: I1008 10:24:16.118210 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6b4cb6945c-kjx6m_d7cda27f-1e6a-4637-b647-54f6a8235ce0/neutron-api/0.log" Oct 08 10:24:16 crc kubenswrapper[4744]: I1008 10:24:16.197384 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6b4cb6945c-kjx6m_d7cda27f-1e6a-4637-b647-54f6a8235ce0/neutron-httpd/0.log" Oct 08 10:24:16 crc kubenswrapper[4744]: I1008 10:24:16.515390 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx_3f2cc21c-f2e1-4775-8563-c794073928f4/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:17 crc kubenswrapper[4744]: I1008 10:24:17.739813 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_63d92921-2a92-4233-8f69-02aa2c2b5984/nova-api-log/0.log" Oct 08 10:24:17 crc kubenswrapper[4744]: I1008 10:24:17.787732 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_3e036c58-fa76-41bb-a16f-77566f7e5533/nova-cell0-conductor-conductor/0.log" Oct 08 10:24:17 crc kubenswrapper[4744]: I1008 10:24:17.991916 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_63d92921-2a92-4233-8f69-02aa2c2b5984/nova-api-api/0.log" Oct 08 10:24:19 crc kubenswrapper[4744]: I1008 10:24:19.019596 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_a197fde0-ffb1-4ead-bdd8-acd2e60bc97f/nova-cell1-novncproxy-novncproxy/0.log" Oct 08 10:24:19 crc kubenswrapper[4744]: I1008 10:24:19.037008 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0b03c126-e043-4a2e-ae42-e9a359622fcd/nova-cell1-conductor-conductor/0.log" Oct 08 10:24:19 crc kubenswrapper[4744]: I1008 10:24:19.436537 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z_43d8aacb-0649-483b-9e07-bde175fd9d52/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:19 crc kubenswrapper[4744]: I1008 10:24:19.656093 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0592f7bc-38a5-43ad-ae49-acdb4f9023f4/nova-metadata-log/0.log" Oct 08 10:24:20 crc kubenswrapper[4744]: I1008 10:24:20.297183 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ee6ac453-8ad5-4c02-adf0-ff88a946dd15/nova-scheduler-scheduler/0.log" Oct 08 10:24:20 crc kubenswrapper[4744]: I1008 10:24:20.629417 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ef3dc460-b269-4e4c-b1db-bfb550b58a7b/mysql-bootstrap/0.log" Oct 08 10:24:20 crc kubenswrapper[4744]: I1008 10:24:20.730482 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ef3dc460-b269-4e4c-b1db-bfb550b58a7b/mysql-bootstrap/0.log" Oct 08 10:24:21 crc kubenswrapper[4744]: I1008 10:24:21.344485 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ef3dc460-b269-4e4c-b1db-bfb550b58a7b/galera/0.log" Oct 08 10:24:21 crc kubenswrapper[4744]: I1008 10:24:21.353078 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d1412b08-0e26-455c-afb5-3f51b2cb3012/mysql-bootstrap/0.log" Oct 08 10:24:21 crc kubenswrapper[4744]: I1008 10:24:21.660346 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d1412b08-0e26-455c-afb5-3f51b2cb3012/galera/0.log" Oct 08 10:24:21 crc kubenswrapper[4744]: I1008 10:24:21.839226 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d1412b08-0e26-455c-afb5-3f51b2cb3012/mysql-bootstrap/0.log" Oct 08 10:24:21 crc kubenswrapper[4744]: I1008 10:24:21.862280 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0592f7bc-38a5-43ad-ae49-acdb4f9023f4/nova-metadata-metadata/0.log" Oct 08 10:24:22 crc kubenswrapper[4744]: I1008 10:24:22.178965 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6fd3ee7f-6102-426d-b482-a624a7871f42/openstackclient/0.log" Oct 08 10:24:22 crc kubenswrapper[4744]: I1008 10:24:22.454611 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kshz8_82e01314-590d-4fe4-9bbb-e5d180ff7093/openstack-network-exporter/0.log" Oct 08 10:24:22 crc kubenswrapper[4744]: I1008 10:24:22.774869 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-mhqr7_ad8606dc-c6d0-4ae7-905e-929d0a82ae8d/ovn-controller/0.log" Oct 08 10:24:22 crc kubenswrapper[4744]: I1008 10:24:22.894771 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ff8dm_fef2319a-5415-4712-a7c6-ad0ab4decff0/ovsdb-server-init/0.log" Oct 08 10:24:23 crc kubenswrapper[4744]: I1008 10:24:23.140433 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ff8dm_fef2319a-5415-4712-a7c6-ad0ab4decff0/ovsdb-server-init/0.log" Oct 08 10:24:23 crc kubenswrapper[4744]: I1008 10:24:23.151519 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ff8dm_fef2319a-5415-4712-a7c6-ad0ab4decff0/ovs-vswitchd/0.log" Oct 08 10:24:23 crc kubenswrapper[4744]: I1008 10:24:23.191556 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ff8dm_fef2319a-5415-4712-a7c6-ad0ab4decff0/ovsdb-server/0.log" Oct 08 10:24:23 crc kubenswrapper[4744]: I1008 10:24:23.655271 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-qslf4_fa1452ca-c8a1-4185-b5ee-9119851d2e20/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:23 crc kubenswrapper[4744]: I1008 10:24:23.743038 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2c0cba37-bef9-4b03-8c6a-43acd36d04ff/openstack-network-exporter/0.log" Oct 08 10:24:23 crc kubenswrapper[4744]: I1008 10:24:23.872916 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2c0cba37-bef9-4b03-8c6a-43acd36d04ff/ovn-northd/0.log" Oct 08 10:24:24 crc kubenswrapper[4744]: I1008 10:24:24.064163 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a38f1196-55fe-462c-b012-be0d5fbe33ee/openstack-network-exporter/0.log" Oct 08 10:24:24 crc kubenswrapper[4744]: I1008 10:24:24.183280 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a38f1196-55fe-462c-b012-be0d5fbe33ee/ovsdbserver-nb/0.log" Oct 08 10:24:24 crc kubenswrapper[4744]: I1008 10:24:24.356430 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_815ecda8-d0f0-47c1-b638-11a5ca9fce02/openstack-network-exporter/0.log" Oct 08 10:24:24 crc kubenswrapper[4744]: I1008 10:24:24.461779 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_815ecda8-d0f0-47c1-b638-11a5ca9fce02/ovsdbserver-sb/0.log" Oct 08 10:24:24 crc kubenswrapper[4744]: I1008 10:24:24.734755 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7f48465966-bh5rl_79381d24-0ce2-46ee-ae85-6fdbcbbae230/placement-api/0.log" Oct 08 10:24:24 crc kubenswrapper[4744]: I1008 10:24:24.843973 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7f48465966-bh5rl_79381d24-0ce2-46ee-ae85-6fdbcbbae230/placement-log/0.log" Oct 08 10:24:25 crc kubenswrapper[4744]: I1008 10:24:25.091226 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8ae09ae5-011d-48f9-ad83-a53bfb91c90c/setup-container/0.log" Oct 08 10:24:25 crc kubenswrapper[4744]: I1008 10:24:25.328503 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8ae09ae5-011d-48f9-ad83-a53bfb91c90c/rabbitmq/0.log" Oct 08 10:24:25 crc kubenswrapper[4744]: I1008 10:24:25.331172 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8ae09ae5-011d-48f9-ad83-a53bfb91c90c/setup-container/0.log" Oct 08 10:24:25 crc kubenswrapper[4744]: I1008 10:24:25.998138 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_19186b52-dfab-4d1f-9ccd-176abc9975ff/setup-container/0.log" Oct 08 10:24:26 crc kubenswrapper[4744]: I1008 10:24:26.454271 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_19186b52-dfab-4d1f-9ccd-176abc9975ff/setup-container/0.log" Oct 08 10:24:26 crc kubenswrapper[4744]: I1008 10:24:26.493550 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_19186b52-dfab-4d1f-9ccd-176abc9975ff/rabbitmq/0.log" Oct 08 10:24:26 crc kubenswrapper[4744]: I1008 10:24:26.700667 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2_446dc944-bd86-4c7a-a9da-bc1193d34d33/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:26 crc kubenswrapper[4744]: I1008 10:24:26.752415 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96_50611401-ad48-4b64-ad34-48a7f2d71a11/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:27 crc kubenswrapper[4744]: I1008 10:24:27.395121 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-jmbgg_77eb2e7f-5360-4589-bb9f-56d6a28fc3b0/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:27 crc kubenswrapper[4744]: I1008 10:24:27.644200 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-f6q5g_ed43f66e-c9f8-4de0-9604-f5a1b993ed03/ssh-known-hosts-edpm-deployment/0.log" Oct 08 10:24:27 crc kubenswrapper[4744]: I1008 10:24:27.959957 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_ba86d64c-cc8a-48b2-b70e-2ca6cf2145be/tempest-tests-tempest-tests-runner/0.log" Oct 08 10:24:28 crc kubenswrapper[4744]: I1008 10:24:28.170944 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_e36a0230-d350-400e-a974-ce78f036dd64/test-operator-logs-container/0.log" Oct 08 10:24:28 crc kubenswrapper[4744]: I1008 10:24:28.605272 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh_4e3fa5bc-5c8b-45de-a173-daf069f2868f/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:24:38 crc kubenswrapper[4744]: I1008 10:24:38.140169 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_4c8e37bd-80bc-4e95-8baf-f571adecb232/memcached/0.log" Oct 08 10:24:48 crc kubenswrapper[4744]: I1008 10:24:48.504954 4744 generic.go:334] "Generic (PLEG): container finished" podID="9c892292-d161-4d31-b101-4f2af00364a1" containerID="2ce2e5bed34ee83a1fba64376cde9515a540d0409a38130a83dde8ee3d95eb0c" exitCode=0 Oct 08 10:24:48 crc kubenswrapper[4744]: I1008 10:24:48.505162 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/crc-debug-2794j" event={"ID":"9c892292-d161-4d31-b101-4f2af00364a1","Type":"ContainerDied","Data":"2ce2e5bed34ee83a1fba64376cde9515a540d0409a38130a83dde8ee3d95eb0c"} Oct 08 10:24:49 crc kubenswrapper[4744]: I1008 10:24:49.621342 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/crc-debug-2794j" Oct 08 10:24:49 crc kubenswrapper[4744]: I1008 10:24:49.656646 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lx88m/crc-debug-2794j"] Oct 08 10:24:49 crc kubenswrapper[4744]: I1008 10:24:49.665242 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lx88m/crc-debug-2794j"] Oct 08 10:24:49 crc kubenswrapper[4744]: I1008 10:24:49.691170 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:24:49 crc kubenswrapper[4744]: I1008 10:24:49.691239 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:24:49 crc kubenswrapper[4744]: I1008 10:24:49.698597 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m97hr\" (UniqueName: \"kubernetes.io/projected/9c892292-d161-4d31-b101-4f2af00364a1-kube-api-access-m97hr\") pod \"9c892292-d161-4d31-b101-4f2af00364a1\" (UID: \"9c892292-d161-4d31-b101-4f2af00364a1\") " Oct 08 10:24:49 crc kubenswrapper[4744]: I1008 10:24:49.698670 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c892292-d161-4d31-b101-4f2af00364a1-host\") pod \"9c892292-d161-4d31-b101-4f2af00364a1\" (UID: \"9c892292-d161-4d31-b101-4f2af00364a1\") " Oct 08 10:24:49 crc kubenswrapper[4744]: I1008 10:24:49.698797 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c892292-d161-4d31-b101-4f2af00364a1-host" (OuterVolumeSpecName: "host") pod "9c892292-d161-4d31-b101-4f2af00364a1" (UID: "9c892292-d161-4d31-b101-4f2af00364a1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 10:24:49 crc kubenswrapper[4744]: I1008 10:24:49.699348 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c892292-d161-4d31-b101-4f2af00364a1-host\") on node \"crc\" DevicePath \"\"" Oct 08 10:24:49 crc kubenswrapper[4744]: I1008 10:24:49.732145 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c892292-d161-4d31-b101-4f2af00364a1-kube-api-access-m97hr" (OuterVolumeSpecName: "kube-api-access-m97hr") pod "9c892292-d161-4d31-b101-4f2af00364a1" (UID: "9c892292-d161-4d31-b101-4f2af00364a1"). InnerVolumeSpecName "kube-api-access-m97hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:24:49 crc kubenswrapper[4744]: I1008 10:24:49.800955 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m97hr\" (UniqueName: \"kubernetes.io/projected/9c892292-d161-4d31-b101-4f2af00364a1-kube-api-access-m97hr\") on node \"crc\" DevicePath \"\"" Oct 08 10:24:50 crc kubenswrapper[4744]: I1008 10:24:50.526045 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9aa69b358fe4e9cd643b64103fdcf3cf36bbc308cad34123bfc6aac85b55ce02" Oct 08 10:24:50 crc kubenswrapper[4744]: I1008 10:24:50.526113 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/crc-debug-2794j" Oct 08 10:24:50 crc kubenswrapper[4744]: I1008 10:24:50.812759 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lx88m/crc-debug-ng6lv"] Oct 08 10:24:50 crc kubenswrapper[4744]: E1008 10:24:50.814669 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c892292-d161-4d31-b101-4f2af00364a1" containerName="container-00" Oct 08 10:24:50 crc kubenswrapper[4744]: I1008 10:24:50.814693 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c892292-d161-4d31-b101-4f2af00364a1" containerName="container-00" Oct 08 10:24:50 crc kubenswrapper[4744]: I1008 10:24:50.814962 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c892292-d161-4d31-b101-4f2af00364a1" containerName="container-00" Oct 08 10:24:50 crc kubenswrapper[4744]: I1008 10:24:50.815576 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/crc-debug-ng6lv" Oct 08 10:24:50 crc kubenswrapper[4744]: I1008 10:24:50.818812 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-lx88m"/"default-dockercfg-mmpt4" Oct 08 10:24:50 crc kubenswrapper[4744]: I1008 10:24:50.921716 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8310330d-90db-4600-a9e1-fe573d8b2bfb-host\") pod \"crc-debug-ng6lv\" (UID: \"8310330d-90db-4600-a9e1-fe573d8b2bfb\") " pod="openshift-must-gather-lx88m/crc-debug-ng6lv" Oct 08 10:24:50 crc kubenswrapper[4744]: I1008 10:24:50.921811 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbkd9\" (UniqueName: \"kubernetes.io/projected/8310330d-90db-4600-a9e1-fe573d8b2bfb-kube-api-access-pbkd9\") pod \"crc-debug-ng6lv\" (UID: \"8310330d-90db-4600-a9e1-fe573d8b2bfb\") " pod="openshift-must-gather-lx88m/crc-debug-ng6lv" Oct 08 10:24:51 crc kubenswrapper[4744]: I1008 10:24:51.023286 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8310330d-90db-4600-a9e1-fe573d8b2bfb-host\") pod \"crc-debug-ng6lv\" (UID: \"8310330d-90db-4600-a9e1-fe573d8b2bfb\") " pod="openshift-must-gather-lx88m/crc-debug-ng6lv" Oct 08 10:24:51 crc kubenswrapper[4744]: I1008 10:24:51.023405 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbkd9\" (UniqueName: \"kubernetes.io/projected/8310330d-90db-4600-a9e1-fe573d8b2bfb-kube-api-access-pbkd9\") pod \"crc-debug-ng6lv\" (UID: \"8310330d-90db-4600-a9e1-fe573d8b2bfb\") " pod="openshift-must-gather-lx88m/crc-debug-ng6lv" Oct 08 10:24:51 crc kubenswrapper[4744]: I1008 10:24:51.023410 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8310330d-90db-4600-a9e1-fe573d8b2bfb-host\") pod \"crc-debug-ng6lv\" (UID: \"8310330d-90db-4600-a9e1-fe573d8b2bfb\") " pod="openshift-must-gather-lx88m/crc-debug-ng6lv" Oct 08 10:24:51 crc kubenswrapper[4744]: I1008 10:24:51.046850 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbkd9\" (UniqueName: \"kubernetes.io/projected/8310330d-90db-4600-a9e1-fe573d8b2bfb-kube-api-access-pbkd9\") pod \"crc-debug-ng6lv\" (UID: \"8310330d-90db-4600-a9e1-fe573d8b2bfb\") " pod="openshift-must-gather-lx88m/crc-debug-ng6lv" Oct 08 10:24:51 crc kubenswrapper[4744]: I1008 10:24:51.134875 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/crc-debug-ng6lv" Oct 08 10:24:51 crc kubenswrapper[4744]: I1008 10:24:51.463833 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c892292-d161-4d31-b101-4f2af00364a1" path="/var/lib/kubelet/pods/9c892292-d161-4d31-b101-4f2af00364a1/volumes" Oct 08 10:24:51 crc kubenswrapper[4744]: I1008 10:24:51.537003 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/crc-debug-ng6lv" event={"ID":"8310330d-90db-4600-a9e1-fe573d8b2bfb","Type":"ContainerStarted","Data":"b46fb2beb2af63fcd68ed6b1414c4c5f32f8f019e1c4533ccb16e3db5515dc32"} Oct 08 10:24:51 crc kubenswrapper[4744]: I1008 10:24:51.537056 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/crc-debug-ng6lv" event={"ID":"8310330d-90db-4600-a9e1-fe573d8b2bfb","Type":"ContainerStarted","Data":"746a3a24a38100f7bdc79426c38231ee16ab5fab468563333e070a0f86fef66e"} Oct 08 10:24:51 crc kubenswrapper[4744]: I1008 10:24:51.555043 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lx88m/crc-debug-ng6lv" podStartSLOduration=1.5550260219999998 podStartE2EDuration="1.555026022s" podCreationTimestamp="2025-10-08 10:24:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:24:51.549342494 +0000 UTC m=+4386.796987733" watchObservedRunningTime="2025-10-08 10:24:51.555026022 +0000 UTC m=+4386.802671261" Oct 08 10:24:52 crc kubenswrapper[4744]: I1008 10:24:52.549412 4744 generic.go:334] "Generic (PLEG): container finished" podID="8310330d-90db-4600-a9e1-fe573d8b2bfb" containerID="b46fb2beb2af63fcd68ed6b1414c4c5f32f8f019e1c4533ccb16e3db5515dc32" exitCode=0 Oct 08 10:24:52 crc kubenswrapper[4744]: I1008 10:24:52.549477 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/crc-debug-ng6lv" event={"ID":"8310330d-90db-4600-a9e1-fe573d8b2bfb","Type":"ContainerDied","Data":"b46fb2beb2af63fcd68ed6b1414c4c5f32f8f019e1c4533ccb16e3db5515dc32"} Oct 08 10:24:53 crc kubenswrapper[4744]: I1008 10:24:53.700291 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/crc-debug-ng6lv" Oct 08 10:24:53 crc kubenswrapper[4744]: I1008 10:24:53.770587 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbkd9\" (UniqueName: \"kubernetes.io/projected/8310330d-90db-4600-a9e1-fe573d8b2bfb-kube-api-access-pbkd9\") pod \"8310330d-90db-4600-a9e1-fe573d8b2bfb\" (UID: \"8310330d-90db-4600-a9e1-fe573d8b2bfb\") " Oct 08 10:24:53 crc kubenswrapper[4744]: I1008 10:24:53.770718 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8310330d-90db-4600-a9e1-fe573d8b2bfb-host\") pod \"8310330d-90db-4600-a9e1-fe573d8b2bfb\" (UID: \"8310330d-90db-4600-a9e1-fe573d8b2bfb\") " Oct 08 10:24:53 crc kubenswrapper[4744]: I1008 10:24:53.771430 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8310330d-90db-4600-a9e1-fe573d8b2bfb-host" (OuterVolumeSpecName: "host") pod "8310330d-90db-4600-a9e1-fe573d8b2bfb" (UID: "8310330d-90db-4600-a9e1-fe573d8b2bfb"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 10:24:53 crc kubenswrapper[4744]: I1008 10:24:53.873003 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8310330d-90db-4600-a9e1-fe573d8b2bfb-host\") on node \"crc\" DevicePath \"\"" Oct 08 10:24:54 crc kubenswrapper[4744]: I1008 10:24:54.198652 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8310330d-90db-4600-a9e1-fe573d8b2bfb-kube-api-access-pbkd9" (OuterVolumeSpecName: "kube-api-access-pbkd9") pod "8310330d-90db-4600-a9e1-fe573d8b2bfb" (UID: "8310330d-90db-4600-a9e1-fe573d8b2bfb"). InnerVolumeSpecName "kube-api-access-pbkd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:24:54 crc kubenswrapper[4744]: I1008 10:24:54.278015 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbkd9\" (UniqueName: \"kubernetes.io/projected/8310330d-90db-4600-a9e1-fe573d8b2bfb-kube-api-access-pbkd9\") on node \"crc\" DevicePath \"\"" Oct 08 10:24:54 crc kubenswrapper[4744]: I1008 10:24:54.573675 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/crc-debug-ng6lv" event={"ID":"8310330d-90db-4600-a9e1-fe573d8b2bfb","Type":"ContainerDied","Data":"746a3a24a38100f7bdc79426c38231ee16ab5fab468563333e070a0f86fef66e"} Oct 08 10:24:54 crc kubenswrapper[4744]: I1008 10:24:54.573716 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="746a3a24a38100f7bdc79426c38231ee16ab5fab468563333e070a0f86fef66e" Oct 08 10:24:54 crc kubenswrapper[4744]: I1008 10:24:54.573719 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/crc-debug-ng6lv" Oct 08 10:24:57 crc kubenswrapper[4744]: I1008 10:24:57.962716 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lx88m/crc-debug-ng6lv"] Oct 08 10:24:57 crc kubenswrapper[4744]: I1008 10:24:57.970040 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lx88m/crc-debug-ng6lv"] Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.141442 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lx88m/crc-debug-jpnqs"] Oct 08 10:24:59 crc kubenswrapper[4744]: E1008 10:24:59.141943 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8310330d-90db-4600-a9e1-fe573d8b2bfb" containerName="container-00" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.141956 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8310330d-90db-4600-a9e1-fe573d8b2bfb" containerName="container-00" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.142192 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8310330d-90db-4600-a9e1-fe573d8b2bfb" containerName="container-00" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.143978 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/crc-debug-jpnqs" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.148224 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-lx88m"/"default-dockercfg-mmpt4" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.178869 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pkr9\" (UniqueName: \"kubernetes.io/projected/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508-kube-api-access-6pkr9\") pod \"crc-debug-jpnqs\" (UID: \"4b2e0f5e-1ecd-4c32-a6aa-238be47ff508\") " pod="openshift-must-gather-lx88m/crc-debug-jpnqs" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.179223 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508-host\") pod \"crc-debug-jpnqs\" (UID: \"4b2e0f5e-1ecd-4c32-a6aa-238be47ff508\") " pod="openshift-must-gather-lx88m/crc-debug-jpnqs" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.281169 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pkr9\" (UniqueName: \"kubernetes.io/projected/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508-kube-api-access-6pkr9\") pod \"crc-debug-jpnqs\" (UID: \"4b2e0f5e-1ecd-4c32-a6aa-238be47ff508\") " pod="openshift-must-gather-lx88m/crc-debug-jpnqs" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.281530 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508-host\") pod \"crc-debug-jpnqs\" (UID: \"4b2e0f5e-1ecd-4c32-a6aa-238be47ff508\") " pod="openshift-must-gather-lx88m/crc-debug-jpnqs" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.281669 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508-host\") pod \"crc-debug-jpnqs\" (UID: \"4b2e0f5e-1ecd-4c32-a6aa-238be47ff508\") " pod="openshift-must-gather-lx88m/crc-debug-jpnqs" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.314552 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pkr9\" (UniqueName: \"kubernetes.io/projected/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508-kube-api-access-6pkr9\") pod \"crc-debug-jpnqs\" (UID: \"4b2e0f5e-1ecd-4c32-a6aa-238be47ff508\") " pod="openshift-must-gather-lx88m/crc-debug-jpnqs" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.464835 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8310330d-90db-4600-a9e1-fe573d8b2bfb" path="/var/lib/kubelet/pods/8310330d-90db-4600-a9e1-fe573d8b2bfb/volumes" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.469955 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/crc-debug-jpnqs" Oct 08 10:24:59 crc kubenswrapper[4744]: I1008 10:24:59.614280 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/crc-debug-jpnqs" event={"ID":"4b2e0f5e-1ecd-4c32-a6aa-238be47ff508","Type":"ContainerStarted","Data":"0a6de599d4a794be9c0cb97cc6571a8bd5149ae2d24ba931edbcb29d21c10799"} Oct 08 10:25:00 crc kubenswrapper[4744]: I1008 10:25:00.622458 4744 generic.go:334] "Generic (PLEG): container finished" podID="4b2e0f5e-1ecd-4c32-a6aa-238be47ff508" containerID="ca986dd51c7945bfed95d9b240eb7690d6859604bd05c7d1b763a5831cda569a" exitCode=0 Oct 08 10:25:00 crc kubenswrapper[4744]: I1008 10:25:00.622601 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/crc-debug-jpnqs" event={"ID":"4b2e0f5e-1ecd-4c32-a6aa-238be47ff508","Type":"ContainerDied","Data":"ca986dd51c7945bfed95d9b240eb7690d6859604bd05c7d1b763a5831cda569a"} Oct 08 10:25:00 crc kubenswrapper[4744]: I1008 10:25:00.657169 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lx88m/crc-debug-jpnqs"] Oct 08 10:25:00 crc kubenswrapper[4744]: I1008 10:25:00.669196 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lx88m/crc-debug-jpnqs"] Oct 08 10:25:02 crc kubenswrapper[4744]: I1008 10:25:02.330119 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/crc-debug-jpnqs" Oct 08 10:25:02 crc kubenswrapper[4744]: I1008 10:25:02.488430 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508-host\") pod \"4b2e0f5e-1ecd-4c32-a6aa-238be47ff508\" (UID: \"4b2e0f5e-1ecd-4c32-a6aa-238be47ff508\") " Oct 08 10:25:02 crc kubenswrapper[4744]: I1008 10:25:02.488546 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pkr9\" (UniqueName: \"kubernetes.io/projected/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508-kube-api-access-6pkr9\") pod \"4b2e0f5e-1ecd-4c32-a6aa-238be47ff508\" (UID: \"4b2e0f5e-1ecd-4c32-a6aa-238be47ff508\") " Oct 08 10:25:02 crc kubenswrapper[4744]: I1008 10:25:02.488537 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508-host" (OuterVolumeSpecName: "host") pod "4b2e0f5e-1ecd-4c32-a6aa-238be47ff508" (UID: "4b2e0f5e-1ecd-4c32-a6aa-238be47ff508"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 10:25:02 crc kubenswrapper[4744]: I1008 10:25:02.489300 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508-host\") on node \"crc\" DevicePath \"\"" Oct 08 10:25:02 crc kubenswrapper[4744]: I1008 10:25:02.495380 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508-kube-api-access-6pkr9" (OuterVolumeSpecName: "kube-api-access-6pkr9") pod "4b2e0f5e-1ecd-4c32-a6aa-238be47ff508" (UID: "4b2e0f5e-1ecd-4c32-a6aa-238be47ff508"). InnerVolumeSpecName "kube-api-access-6pkr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:25:02 crc kubenswrapper[4744]: I1008 10:25:02.590835 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pkr9\" (UniqueName: \"kubernetes.io/projected/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508-kube-api-access-6pkr9\") on node \"crc\" DevicePath \"\"" Oct 08 10:25:02 crc kubenswrapper[4744]: I1008 10:25:02.640781 4744 scope.go:117] "RemoveContainer" containerID="ca986dd51c7945bfed95d9b240eb7690d6859604bd05c7d1b763a5831cda569a" Oct 08 10:25:02 crc kubenswrapper[4744]: I1008 10:25:02.640818 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/crc-debug-jpnqs" Oct 08 10:25:03 crc kubenswrapper[4744]: I1008 10:25:03.023621 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-nzjx2_6d6f4234-2b91-4399-8aa2-f1064fde501c/kube-rbac-proxy/0.log" Oct 08 10:25:03 crc kubenswrapper[4744]: I1008 10:25:03.108611 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-nzjx2_6d6f4234-2b91-4399-8aa2-f1064fde501c/manager/0.log" Oct 08 10:25:03 crc kubenswrapper[4744]: I1008 10:25:03.292700 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-csxmb_8a43b240-9fe3-4c10-afa9-7e78777e76d5/kube-rbac-proxy/0.log" Oct 08 10:25:03 crc kubenswrapper[4744]: I1008 10:25:03.341775 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-csxmb_8a43b240-9fe3-4c10-afa9-7e78777e76d5/manager/0.log" Oct 08 10:25:03 crc kubenswrapper[4744]: I1008 10:25:03.464770 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b2e0f5e-1ecd-4c32-a6aa-238be47ff508" path="/var/lib/kubelet/pods/4b2e0f5e-1ecd-4c32-a6aa-238be47ff508/volumes" Oct 08 10:25:03 crc kubenswrapper[4744]: I1008 10:25:03.984010 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-b9ntl_254b5ec4-bc16-4e8a-a61b-8a4a588dd629/kube-rbac-proxy/0.log" Oct 08 10:25:03 crc kubenswrapper[4744]: I1008 10:25:03.986789 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/util/0.log" Oct 08 10:25:03 crc kubenswrapper[4744]: I1008 10:25:03.995385 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-b9ntl_254b5ec4-bc16-4e8a-a61b-8a4a588dd629/manager/0.log" Oct 08 10:25:04 crc kubenswrapper[4744]: I1008 10:25:04.256026 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/pull/0.log" Oct 08 10:25:04 crc kubenswrapper[4744]: I1008 10:25:04.298149 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/util/0.log" Oct 08 10:25:04 crc kubenswrapper[4744]: I1008 10:25:04.430462 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/pull/0.log" Oct 08 10:25:04 crc kubenswrapper[4744]: I1008 10:25:04.597353 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/pull/0.log" Oct 08 10:25:04 crc kubenswrapper[4744]: I1008 10:25:04.625119 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/util/0.log" Oct 08 10:25:04 crc kubenswrapper[4744]: I1008 10:25:04.749480 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/extract/0.log" Oct 08 10:25:04 crc kubenswrapper[4744]: I1008 10:25:04.885400 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-f9zmb_44cd0502-66fb-45ef-888a-da6273df7056/kube-rbac-proxy/0.log" Oct 08 10:25:05 crc kubenswrapper[4744]: I1008 10:25:05.082274 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-m9prk_622e796c-d598-4721-89bd-c0cb3f83a1ad/kube-rbac-proxy/0.log" Oct 08 10:25:05 crc kubenswrapper[4744]: I1008 10:25:05.109995 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-f9zmb_44cd0502-66fb-45ef-888a-da6273df7056/manager/0.log" Oct 08 10:25:05 crc kubenswrapper[4744]: I1008 10:25:05.183299 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-m9prk_622e796c-d598-4721-89bd-c0cb3f83a1ad/manager/0.log" Oct 08 10:25:05 crc kubenswrapper[4744]: I1008 10:25:05.438861 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-hsmrl_18a79d6e-9afe-43b5-8cf1-b5a991d2fd49/manager/0.log" Oct 08 10:25:05 crc kubenswrapper[4744]: I1008 10:25:05.461609 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-hsmrl_18a79d6e-9afe-43b5-8cf1-b5a991d2fd49/kube-rbac-proxy/0.log" Oct 08 10:25:05 crc kubenswrapper[4744]: I1008 10:25:05.687029 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-tbx59_b2bb8369-17dd-4f3b-8e8d-4af7895d892c/kube-rbac-proxy/0.log" Oct 08 10:25:05 crc kubenswrapper[4744]: I1008 10:25:05.762618 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-q9hcb_6126b107-8c51-4467-9c40-e2c4b268cec5/kube-rbac-proxy/0.log" Oct 08 10:25:05 crc kubenswrapper[4744]: I1008 10:25:05.827675 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-tbx59_b2bb8369-17dd-4f3b-8e8d-4af7895d892c/manager/0.log" Oct 08 10:25:06 crc kubenswrapper[4744]: I1008 10:25:06.012563 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-q9hcb_6126b107-8c51-4467-9c40-e2c4b268cec5/manager/0.log" Oct 08 10:25:06 crc kubenswrapper[4744]: I1008 10:25:06.042392 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-xjxx8_72a5635b-6920-4c2b-b91c-7621bc933959/kube-rbac-proxy/0.log" Oct 08 10:25:06 crc kubenswrapper[4744]: I1008 10:25:06.152143 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-xjxx8_72a5635b-6920-4c2b-b91c-7621bc933959/manager/0.log" Oct 08 10:25:06 crc kubenswrapper[4744]: I1008 10:25:06.313050 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-lj9pr_58decb1a-987e-4366-9bf8-ad4bf73e5969/manager/0.log" Oct 08 10:25:06 crc kubenswrapper[4744]: I1008 10:25:06.319997 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-lj9pr_58decb1a-987e-4366-9bf8-ad4bf73e5969/kube-rbac-proxy/0.log" Oct 08 10:25:06 crc kubenswrapper[4744]: I1008 10:25:06.554441 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-4xvzd_5fae63cb-c599-447b-bafb-2ed6bad836d3/kube-rbac-proxy/0.log" Oct 08 10:25:06 crc kubenswrapper[4744]: I1008 10:25:06.589207 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-xl4hw_48158343-3842-44df-86c8-3a2f0e06a09c/kube-rbac-proxy/0.log" Oct 08 10:25:06 crc kubenswrapper[4744]: I1008 10:25:06.631088 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-4xvzd_5fae63cb-c599-447b-bafb-2ed6bad836d3/manager/0.log" Oct 08 10:25:06 crc kubenswrapper[4744]: I1008 10:25:06.709583 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-xl4hw_48158343-3842-44df-86c8-3a2f0e06a09c/manager/0.log" Oct 08 10:25:06 crc kubenswrapper[4744]: I1008 10:25:06.780689 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-5vz2c_2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34/kube-rbac-proxy/0.log" Oct 08 10:25:06 crc kubenswrapper[4744]: I1008 10:25:06.909708 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-5vz2c_2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34/manager/0.log" Oct 08 10:25:06 crc kubenswrapper[4744]: I1008 10:25:06.937325 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-crqx2_3eeb379f-9005-4908-ac9e-6dfb06405fb2/kube-rbac-proxy/0.log" Oct 08 10:25:07 crc kubenswrapper[4744]: I1008 10:25:07.030092 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-crqx2_3eeb379f-9005-4908-ac9e-6dfb06405fb2/manager/0.log" Oct 08 10:25:07 crc kubenswrapper[4744]: I1008 10:25:07.118360 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-747747dfccwc6fw_c5096473-ff21-474e-b7fc-03196e4ee5a8/kube-rbac-proxy/0.log" Oct 08 10:25:07 crc kubenswrapper[4744]: I1008 10:25:07.164253 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-747747dfccwc6fw_c5096473-ff21-474e-b7fc-03196e4ee5a8/manager/0.log" Oct 08 10:25:07 crc kubenswrapper[4744]: I1008 10:25:07.191466 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8bc6b8f5b-7q7gz_6fdbcdf9-cab4-4d8a-9270-2a121c83dc04/kube-rbac-proxy/0.log" Oct 08 10:25:07 crc kubenswrapper[4744]: I1008 10:25:07.455775 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-55f65988b-bjktr_b575f955-1e82-4afa-a84b-842c2ba7e47b/kube-rbac-proxy/0.log" Oct 08 10:25:07 crc kubenswrapper[4744]: I1008 10:25:07.594073 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-55f65988b-bjktr_b575f955-1e82-4afa-a84b-842c2ba7e47b/operator/0.log" Oct 08 10:25:07 crc kubenswrapper[4744]: I1008 10:25:07.827283 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79db49b9fb-qppfc_e991d608-9aa2-4028-8080-a6bd7dca66c8/kube-rbac-proxy/0.log" Oct 08 10:25:07 crc kubenswrapper[4744]: I1008 10:25:07.951757 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tc4hq_7f7736f3-f7de-4ff0-ae46-f01bf32693f5/registry-server/0.log" Oct 08 10:25:07 crc kubenswrapper[4744]: I1008 10:25:07.980970 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79db49b9fb-qppfc_e991d608-9aa2-4028-8080-a6bd7dca66c8/manager/0.log" Oct 08 10:25:08 crc kubenswrapper[4744]: I1008 10:25:08.183833 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-lbvtv_3a897464-5474-43db-99d5-35f691f64220/kube-rbac-proxy/0.log" Oct 08 10:25:08 crc kubenswrapper[4744]: I1008 10:25:08.271730 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-lbvtv_3a897464-5474-43db-99d5-35f691f64220/manager/0.log" Oct 08 10:25:08 crc kubenswrapper[4744]: I1008 10:25:08.407528 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-ztld4_5bf79b31-d3bc-4422-9302-ac7a55a1d95d/operator/0.log" Oct 08 10:25:08 crc kubenswrapper[4744]: I1008 10:25:08.492975 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-frfr7_663cead5-7b5e-4940-94f9-5974f0424eda/manager/0.log" Oct 08 10:25:08 crc kubenswrapper[4744]: I1008 10:25:08.580568 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-frfr7_663cead5-7b5e-4940-94f9-5974f0424eda/kube-rbac-proxy/0.log" Oct 08 10:25:08 crc kubenswrapper[4744]: I1008 10:25:08.721842 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8bc6b8f5b-7q7gz_6fdbcdf9-cab4-4d8a-9270-2a121c83dc04/manager/0.log" Oct 08 10:25:08 crc kubenswrapper[4744]: I1008 10:25:08.757024 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76796d4c6b-gv4tw_f3456365-6f0d-4087-b753-40292c80bf12/kube-rbac-proxy/0.log" Oct 08 10:25:08 crc kubenswrapper[4744]: I1008 10:25:08.885672 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76796d4c6b-gv4tw_f3456365-6f0d-4087-b753-40292c80bf12/manager/0.log" Oct 08 10:25:08 crc kubenswrapper[4744]: I1008 10:25:08.961213 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56c698c775-bf9cm_12b5913f-74df-4ba6-8aa8-af5989327ddd/kube-rbac-proxy/0.log" Oct 08 10:25:09 crc kubenswrapper[4744]: I1008 10:25:09.003042 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56c698c775-bf9cm_12b5913f-74df-4ba6-8aa8-af5989327ddd/manager/0.log" Oct 08 10:25:09 crc kubenswrapper[4744]: I1008 10:25:09.129478 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7794bc6bd-q884x_60f0c455-812d-40fe-b1b6-5aa75c34753c/kube-rbac-proxy/0.log" Oct 08 10:25:09 crc kubenswrapper[4744]: I1008 10:25:09.146011 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7794bc6bd-q884x_60f0c455-812d-40fe-b1b6-5aa75c34753c/manager/0.log" Oct 08 10:25:19 crc kubenswrapper[4744]: I1008 10:25:19.690238 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:25:19 crc kubenswrapper[4744]: I1008 10:25:19.690831 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:25:25 crc kubenswrapper[4744]: I1008 10:25:25.918872 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5nnfc_81cf0144-e970-4fe2-8663-15b006dc1b91/control-plane-machine-set-operator/0.log" Oct 08 10:25:26 crc kubenswrapper[4744]: I1008 10:25:26.122221 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-b5q8z_a7221db0-3f54-4e07-95ec-56c4fb95aecc/kube-rbac-proxy/0.log" Oct 08 10:25:26 crc kubenswrapper[4744]: I1008 10:25:26.139269 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-b5q8z_a7221db0-3f54-4e07-95ec-56c4fb95aecc/machine-api-operator/0.log" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.569485 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r966p"] Oct 08 10:25:38 crc kubenswrapper[4744]: E1008 10:25:38.571624 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b2e0f5e-1ecd-4c32-a6aa-238be47ff508" containerName="container-00" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.571742 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b2e0f5e-1ecd-4c32-a6aa-238be47ff508" containerName="container-00" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.572135 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b2e0f5e-1ecd-4c32-a6aa-238be47ff508" containerName="container-00" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.577576 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.665906 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r966p"] Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.738635 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2821f947-c8b3-429f-995c-89dca9197d07-catalog-content\") pod \"community-operators-r966p\" (UID: \"2821f947-c8b3-429f-995c-89dca9197d07\") " pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.738875 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2821f947-c8b3-429f-995c-89dca9197d07-utilities\") pod \"community-operators-r966p\" (UID: \"2821f947-c8b3-429f-995c-89dca9197d07\") " pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.739278 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7r97\" (UniqueName: \"kubernetes.io/projected/2821f947-c8b3-429f-995c-89dca9197d07-kube-api-access-q7r97\") pod \"community-operators-r966p\" (UID: \"2821f947-c8b3-429f-995c-89dca9197d07\") " pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.841420 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7r97\" (UniqueName: \"kubernetes.io/projected/2821f947-c8b3-429f-995c-89dca9197d07-kube-api-access-q7r97\") pod \"community-operators-r966p\" (UID: \"2821f947-c8b3-429f-995c-89dca9197d07\") " pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.841556 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2821f947-c8b3-429f-995c-89dca9197d07-catalog-content\") pod \"community-operators-r966p\" (UID: \"2821f947-c8b3-429f-995c-89dca9197d07\") " pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.841630 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2821f947-c8b3-429f-995c-89dca9197d07-utilities\") pod \"community-operators-r966p\" (UID: \"2821f947-c8b3-429f-995c-89dca9197d07\") " pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.843308 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2821f947-c8b3-429f-995c-89dca9197d07-utilities\") pod \"community-operators-r966p\" (UID: \"2821f947-c8b3-429f-995c-89dca9197d07\") " pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.843452 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2821f947-c8b3-429f-995c-89dca9197d07-catalog-content\") pod \"community-operators-r966p\" (UID: \"2821f947-c8b3-429f-995c-89dca9197d07\") " pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.878279 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7r97\" (UniqueName: \"kubernetes.io/projected/2821f947-c8b3-429f-995c-89dca9197d07-kube-api-access-q7r97\") pod \"community-operators-r966p\" (UID: \"2821f947-c8b3-429f-995c-89dca9197d07\") " pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:38 crc kubenswrapper[4744]: I1008 10:25:38.901185 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:40 crc kubenswrapper[4744]: I1008 10:25:40.270769 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r966p"] Oct 08 10:25:40 crc kubenswrapper[4744]: I1008 10:25:40.961078 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-55k9q"] Oct 08 10:25:40 crc kubenswrapper[4744]: I1008 10:25:40.963468 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.024628 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-55k9q"] Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.044314 4744 generic.go:334] "Generic (PLEG): container finished" podID="2821f947-c8b3-429f-995c-89dca9197d07" containerID="bf0ac8a1833dc99412c98f7cea16b85f42b80ebf156f3d698b81a3de2625160a" exitCode=0 Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.044382 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r966p" event={"ID":"2821f947-c8b3-429f-995c-89dca9197d07","Type":"ContainerDied","Data":"bf0ac8a1833dc99412c98f7cea16b85f42b80ebf156f3d698b81a3de2625160a"} Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.044411 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r966p" event={"ID":"2821f947-c8b3-429f-995c-89dca9197d07","Type":"ContainerStarted","Data":"db053a2bf4570537b5d815897928c26e44ae5c236a9731a74a29643fca1d58aa"} Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.086616 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvf5z\" (UniqueName: \"kubernetes.io/projected/2af19560-76bd-4524-b537-ac61cfe6730a-kube-api-access-tvf5z\") pod \"certified-operators-55k9q\" (UID: \"2af19560-76bd-4524-b537-ac61cfe6730a\") " pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.086742 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2af19560-76bd-4524-b537-ac61cfe6730a-utilities\") pod \"certified-operators-55k9q\" (UID: \"2af19560-76bd-4524-b537-ac61cfe6730a\") " pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.086806 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2af19560-76bd-4524-b537-ac61cfe6730a-catalog-content\") pod \"certified-operators-55k9q\" (UID: \"2af19560-76bd-4524-b537-ac61cfe6730a\") " pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.188723 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2af19560-76bd-4524-b537-ac61cfe6730a-utilities\") pod \"certified-operators-55k9q\" (UID: \"2af19560-76bd-4524-b537-ac61cfe6730a\") " pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.188838 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2af19560-76bd-4524-b537-ac61cfe6730a-catalog-content\") pod \"certified-operators-55k9q\" (UID: \"2af19560-76bd-4524-b537-ac61cfe6730a\") " pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.188944 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvf5z\" (UniqueName: \"kubernetes.io/projected/2af19560-76bd-4524-b537-ac61cfe6730a-kube-api-access-tvf5z\") pod \"certified-operators-55k9q\" (UID: \"2af19560-76bd-4524-b537-ac61cfe6730a\") " pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.189887 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2af19560-76bd-4524-b537-ac61cfe6730a-utilities\") pod \"certified-operators-55k9q\" (UID: \"2af19560-76bd-4524-b537-ac61cfe6730a\") " pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.190183 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2af19560-76bd-4524-b537-ac61cfe6730a-catalog-content\") pod \"certified-operators-55k9q\" (UID: \"2af19560-76bd-4524-b537-ac61cfe6730a\") " pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.208159 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvf5z\" (UniqueName: \"kubernetes.io/projected/2af19560-76bd-4524-b537-ac61cfe6730a-kube-api-access-tvf5z\") pod \"certified-operators-55k9q\" (UID: \"2af19560-76bd-4524-b537-ac61cfe6730a\") " pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.286639 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:41 crc kubenswrapper[4744]: I1008 10:25:41.820098 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-55k9q"] Oct 08 10:25:42 crc kubenswrapper[4744]: I1008 10:25:42.053938 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55k9q" event={"ID":"2af19560-76bd-4524-b537-ac61cfe6730a","Type":"ContainerStarted","Data":"0d60f1bcefa4348ebdac4fc587176672e018e7cd67101fdb50d0099a929b766a"} Oct 08 10:25:42 crc kubenswrapper[4744]: I1008 10:25:42.169101 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-87br5_5fde9304-c103-4e8a-96bc-d47a0dea1635/cert-manager-controller/0.log" Oct 08 10:25:42 crc kubenswrapper[4744]: I1008 10:25:42.334487 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-j5pzr_b2fe7e00-2fdb-4186-81db-c211ef88b772/cert-manager-cainjector/0.log" Oct 08 10:25:42 crc kubenswrapper[4744]: I1008 10:25:42.401661 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-4478x_39d1521a-c337-4259-bc31-0bf6357ce325/cert-manager-webhook/0.log" Oct 08 10:25:43 crc kubenswrapper[4744]: I1008 10:25:43.066865 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r966p" event={"ID":"2821f947-c8b3-429f-995c-89dca9197d07","Type":"ContainerStarted","Data":"0fb326ea64e2e8f005a3805d8b90b997a09f46484354c3d54dbd49097f5580f2"} Oct 08 10:25:43 crc kubenswrapper[4744]: I1008 10:25:43.069063 4744 generic.go:334] "Generic (PLEG): container finished" podID="2af19560-76bd-4524-b537-ac61cfe6730a" containerID="b771b1d222595d28da8857b023759b9c17db5324da03a3a1d813c0a636d9cdb5" exitCode=0 Oct 08 10:25:43 crc kubenswrapper[4744]: I1008 10:25:43.069092 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55k9q" event={"ID":"2af19560-76bd-4524-b537-ac61cfe6730a","Type":"ContainerDied","Data":"b771b1d222595d28da8857b023759b9c17db5324da03a3a1d813c0a636d9cdb5"} Oct 08 10:25:45 crc kubenswrapper[4744]: I1008 10:25:45.091210 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55k9q" event={"ID":"2af19560-76bd-4524-b537-ac61cfe6730a","Type":"ContainerStarted","Data":"bd35ec856b3c7684511cca7da74734bbfd43995e940acc887102bad23e12b4fc"} Oct 08 10:25:46 crc kubenswrapper[4744]: I1008 10:25:46.101448 4744 generic.go:334] "Generic (PLEG): container finished" podID="2821f947-c8b3-429f-995c-89dca9197d07" containerID="0fb326ea64e2e8f005a3805d8b90b997a09f46484354c3d54dbd49097f5580f2" exitCode=0 Oct 08 10:25:46 crc kubenswrapper[4744]: I1008 10:25:46.101511 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r966p" event={"ID":"2821f947-c8b3-429f-995c-89dca9197d07","Type":"ContainerDied","Data":"0fb326ea64e2e8f005a3805d8b90b997a09f46484354c3d54dbd49097f5580f2"} Oct 08 10:25:47 crc kubenswrapper[4744]: I1008 10:25:47.112134 4744 generic.go:334] "Generic (PLEG): container finished" podID="2af19560-76bd-4524-b537-ac61cfe6730a" containerID="bd35ec856b3c7684511cca7da74734bbfd43995e940acc887102bad23e12b4fc" exitCode=0 Oct 08 10:25:47 crc kubenswrapper[4744]: I1008 10:25:47.112210 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55k9q" event={"ID":"2af19560-76bd-4524-b537-ac61cfe6730a","Type":"ContainerDied","Data":"bd35ec856b3c7684511cca7da74734bbfd43995e940acc887102bad23e12b4fc"} Oct 08 10:25:47 crc kubenswrapper[4744]: I1008 10:25:47.115660 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r966p" event={"ID":"2821f947-c8b3-429f-995c-89dca9197d07","Type":"ContainerStarted","Data":"ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36"} Oct 08 10:25:47 crc kubenswrapper[4744]: I1008 10:25:47.162847 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r966p" podStartSLOduration=3.5986470280000002 podStartE2EDuration="9.162822734s" podCreationTimestamp="2025-10-08 10:25:38 +0000 UTC" firstStartedPulling="2025-10-08 10:25:41.046616948 +0000 UTC m=+4436.294262187" lastFinishedPulling="2025-10-08 10:25:46.610792654 +0000 UTC m=+4441.858437893" observedRunningTime="2025-10-08 10:25:47.156044665 +0000 UTC m=+4442.403689924" watchObservedRunningTime="2025-10-08 10:25:47.162822734 +0000 UTC m=+4442.410467973" Oct 08 10:25:48 crc kubenswrapper[4744]: I1008 10:25:48.127715 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55k9q" event={"ID":"2af19560-76bd-4524-b537-ac61cfe6730a","Type":"ContainerStarted","Data":"ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea"} Oct 08 10:25:48 crc kubenswrapper[4744]: I1008 10:25:48.902505 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:48 crc kubenswrapper[4744]: I1008 10:25:48.902826 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:49 crc kubenswrapper[4744]: I1008 10:25:49.690456 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:25:49 crc kubenswrapper[4744]: I1008 10:25:49.690537 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:25:49 crc kubenswrapper[4744]: I1008 10:25:49.690588 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 10:25:49 crc kubenswrapper[4744]: I1008 10:25:49.692394 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 10:25:49 crc kubenswrapper[4744]: I1008 10:25:49.692475 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" gracePeriod=600 Oct 08 10:25:49 crc kubenswrapper[4744]: E1008 10:25:49.834029 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:25:49 crc kubenswrapper[4744]: I1008 10:25:49.953177 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-r966p" podUID="2821f947-c8b3-429f-995c-89dca9197d07" containerName="registry-server" probeResult="failure" output=< Oct 08 10:25:49 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 10:25:49 crc kubenswrapper[4744]: > Oct 08 10:25:50 crc kubenswrapper[4744]: I1008 10:25:50.157008 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" exitCode=0 Oct 08 10:25:50 crc kubenswrapper[4744]: I1008 10:25:50.157064 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab"} Oct 08 10:25:50 crc kubenswrapper[4744]: I1008 10:25:50.157106 4744 scope.go:117] "RemoveContainer" containerID="ca8fafcf7cc39e7299b68be5cb7e1e887b65f71741164b2f539597d73f629078" Oct 08 10:25:50 crc kubenswrapper[4744]: I1008 10:25:50.157889 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:25:50 crc kubenswrapper[4744]: E1008 10:25:50.158169 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:25:50 crc kubenswrapper[4744]: I1008 10:25:50.177014 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-55k9q" podStartSLOduration=5.678448463 podStartE2EDuration="10.176994647s" podCreationTimestamp="2025-10-08 10:25:40 +0000 UTC" firstStartedPulling="2025-10-08 10:25:43.070965943 +0000 UTC m=+4438.318611182" lastFinishedPulling="2025-10-08 10:25:47.569512127 +0000 UTC m=+4442.817157366" observedRunningTime="2025-10-08 10:25:48.155757087 +0000 UTC m=+4443.403402326" watchObservedRunningTime="2025-10-08 10:25:50.176994647 +0000 UTC m=+4445.424639906" Oct 08 10:25:51 crc kubenswrapper[4744]: I1008 10:25:51.288198 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:51 crc kubenswrapper[4744]: I1008 10:25:51.288530 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:51 crc kubenswrapper[4744]: I1008 10:25:51.746342 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:52 crc kubenswrapper[4744]: I1008 10:25:52.220289 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:53 crc kubenswrapper[4744]: I1008 10:25:53.161543 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-55k9q"] Oct 08 10:25:54 crc kubenswrapper[4744]: I1008 10:25:54.195534 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-55k9q" podUID="2af19560-76bd-4524-b537-ac61cfe6730a" containerName="registry-server" containerID="cri-o://ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea" gracePeriod=2 Oct 08 10:25:54 crc kubenswrapper[4744]: I1008 10:25:54.962664 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.037537 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-hrjff_2c9804b5-0729-4059-bb00-35fd1d355b3c/nmstate-console-plugin/0.log" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.114932 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvf5z\" (UniqueName: \"kubernetes.io/projected/2af19560-76bd-4524-b537-ac61cfe6730a-kube-api-access-tvf5z\") pod \"2af19560-76bd-4524-b537-ac61cfe6730a\" (UID: \"2af19560-76bd-4524-b537-ac61cfe6730a\") " Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.114985 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2af19560-76bd-4524-b537-ac61cfe6730a-catalog-content\") pod \"2af19560-76bd-4524-b537-ac61cfe6730a\" (UID: \"2af19560-76bd-4524-b537-ac61cfe6730a\") " Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.115024 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2af19560-76bd-4524-b537-ac61cfe6730a-utilities\") pod \"2af19560-76bd-4524-b537-ac61cfe6730a\" (UID: \"2af19560-76bd-4524-b537-ac61cfe6730a\") " Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.116095 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2af19560-76bd-4524-b537-ac61cfe6730a-utilities" (OuterVolumeSpecName: "utilities") pod "2af19560-76bd-4524-b537-ac61cfe6730a" (UID: "2af19560-76bd-4524-b537-ac61cfe6730a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.132068 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2af19560-76bd-4524-b537-ac61cfe6730a-kube-api-access-tvf5z" (OuterVolumeSpecName: "kube-api-access-tvf5z") pod "2af19560-76bd-4524-b537-ac61cfe6730a" (UID: "2af19560-76bd-4524-b537-ac61cfe6730a"). InnerVolumeSpecName "kube-api-access-tvf5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.194422 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2af19560-76bd-4524-b537-ac61cfe6730a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2af19560-76bd-4524-b537-ac61cfe6730a" (UID: "2af19560-76bd-4524-b537-ac61cfe6730a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.205283 4744 generic.go:334] "Generic (PLEG): container finished" podID="2af19560-76bd-4524-b537-ac61cfe6730a" containerID="ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea" exitCode=0 Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.205334 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55k9q" event={"ID":"2af19560-76bd-4524-b537-ac61cfe6730a","Type":"ContainerDied","Data":"ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea"} Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.205386 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-55k9q" event={"ID":"2af19560-76bd-4524-b537-ac61cfe6730a","Type":"ContainerDied","Data":"0d60f1bcefa4348ebdac4fc587176672e018e7cd67101fdb50d0099a929b766a"} Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.205410 4744 scope.go:117] "RemoveContainer" containerID="ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.206406 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-55k9q" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.217398 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvf5z\" (UniqueName: \"kubernetes.io/projected/2af19560-76bd-4524-b537-ac61cfe6730a-kube-api-access-tvf5z\") on node \"crc\" DevicePath \"\"" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.217426 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2af19560-76bd-4524-b537-ac61cfe6730a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.217437 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2af19560-76bd-4524-b537-ac61cfe6730a-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.224117 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-tk2k5_8b5fc549-4718-4e89-82a9-b39e93a2220d/nmstate-handler/0.log" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.226218 4744 scope.go:117] "RemoveContainer" containerID="bd35ec856b3c7684511cca7da74734bbfd43995e940acc887102bad23e12b4fc" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.242938 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-55k9q"] Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.251689 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-55k9q"] Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.265521 4744 scope.go:117] "RemoveContainer" containerID="b771b1d222595d28da8857b023759b9c17db5324da03a3a1d813c0a636d9cdb5" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.311157 4744 scope.go:117] "RemoveContainer" containerID="ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea" Oct 08 10:25:55 crc kubenswrapper[4744]: E1008 10:25:55.311763 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea\": container with ID starting with ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea not found: ID does not exist" containerID="ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.311801 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea"} err="failed to get container status \"ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea\": rpc error: code = NotFound desc = could not find container \"ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea\": container with ID starting with ef4fdfa08a0f4878ff8f1fe4495a2181b0f480790741516547515e5cccbb6aea not found: ID does not exist" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.311826 4744 scope.go:117] "RemoveContainer" containerID="bd35ec856b3c7684511cca7da74734bbfd43995e940acc887102bad23e12b4fc" Oct 08 10:25:55 crc kubenswrapper[4744]: E1008 10:25:55.312081 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd35ec856b3c7684511cca7da74734bbfd43995e940acc887102bad23e12b4fc\": container with ID starting with bd35ec856b3c7684511cca7da74734bbfd43995e940acc887102bad23e12b4fc not found: ID does not exist" containerID="bd35ec856b3c7684511cca7da74734bbfd43995e940acc887102bad23e12b4fc" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.312104 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd35ec856b3c7684511cca7da74734bbfd43995e940acc887102bad23e12b4fc"} err="failed to get container status \"bd35ec856b3c7684511cca7da74734bbfd43995e940acc887102bad23e12b4fc\": rpc error: code = NotFound desc = could not find container \"bd35ec856b3c7684511cca7da74734bbfd43995e940acc887102bad23e12b4fc\": container with ID starting with bd35ec856b3c7684511cca7da74734bbfd43995e940acc887102bad23e12b4fc not found: ID does not exist" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.312117 4744 scope.go:117] "RemoveContainer" containerID="b771b1d222595d28da8857b023759b9c17db5324da03a3a1d813c0a636d9cdb5" Oct 08 10:25:55 crc kubenswrapper[4744]: E1008 10:25:55.312416 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b771b1d222595d28da8857b023759b9c17db5324da03a3a1d813c0a636d9cdb5\": container with ID starting with b771b1d222595d28da8857b023759b9c17db5324da03a3a1d813c0a636d9cdb5 not found: ID does not exist" containerID="b771b1d222595d28da8857b023759b9c17db5324da03a3a1d813c0a636d9cdb5" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.312439 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b771b1d222595d28da8857b023759b9c17db5324da03a3a1d813c0a636d9cdb5"} err="failed to get container status \"b771b1d222595d28da8857b023759b9c17db5324da03a3a1d813c0a636d9cdb5\": rpc error: code = NotFound desc = could not find container \"b771b1d222595d28da8857b023759b9c17db5324da03a3a1d813c0a636d9cdb5\": container with ID starting with b771b1d222595d28da8857b023759b9c17db5324da03a3a1d813c0a636d9cdb5 not found: ID does not exist" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.409570 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-9qfcx_db58dfb9-4fd9-4896-add0-116a1ed9079d/kube-rbac-proxy/0.log" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.463122 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2af19560-76bd-4524-b537-ac61cfe6730a" path="/var/lib/kubelet/pods/2af19560-76bd-4524-b537-ac61cfe6730a/volumes" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.493560 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-9qfcx_db58dfb9-4fd9-4896-add0-116a1ed9079d/nmstate-metrics/0.log" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.613172 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-nmnhx_1f1d4ca0-878e-4211-a40a-37f16085fe47/nmstate-operator/0.log" Oct 08 10:25:55 crc kubenswrapper[4744]: I1008 10:25:55.740024 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-pnz4l_89b615a3-0a5d-483c-8d41-2dee267f1686/nmstate-webhook/0.log" Oct 08 10:25:59 crc kubenswrapper[4744]: I1008 10:25:59.252052 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:59 crc kubenswrapper[4744]: I1008 10:25:59.335000 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r966p" Oct 08 10:25:59 crc kubenswrapper[4744]: I1008 10:25:59.491527 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r966p"] Oct 08 10:26:01 crc kubenswrapper[4744]: I1008 10:26:01.258692 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r966p" podUID="2821f947-c8b3-429f-995c-89dca9197d07" containerName="registry-server" containerID="cri-o://ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36" gracePeriod=2 Oct 08 10:26:01 crc kubenswrapper[4744]: I1008 10:26:01.756033 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r966p" Oct 08 10:26:01 crc kubenswrapper[4744]: I1008 10:26:01.870247 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7r97\" (UniqueName: \"kubernetes.io/projected/2821f947-c8b3-429f-995c-89dca9197d07-kube-api-access-q7r97\") pod \"2821f947-c8b3-429f-995c-89dca9197d07\" (UID: \"2821f947-c8b3-429f-995c-89dca9197d07\") " Oct 08 10:26:01 crc kubenswrapper[4744]: I1008 10:26:01.870430 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2821f947-c8b3-429f-995c-89dca9197d07-catalog-content\") pod \"2821f947-c8b3-429f-995c-89dca9197d07\" (UID: \"2821f947-c8b3-429f-995c-89dca9197d07\") " Oct 08 10:26:01 crc kubenswrapper[4744]: I1008 10:26:01.870526 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2821f947-c8b3-429f-995c-89dca9197d07-utilities\") pod \"2821f947-c8b3-429f-995c-89dca9197d07\" (UID: \"2821f947-c8b3-429f-995c-89dca9197d07\") " Oct 08 10:26:01 crc kubenswrapper[4744]: I1008 10:26:01.871591 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2821f947-c8b3-429f-995c-89dca9197d07-utilities" (OuterVolumeSpecName: "utilities") pod "2821f947-c8b3-429f-995c-89dca9197d07" (UID: "2821f947-c8b3-429f-995c-89dca9197d07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:26:01 crc kubenswrapper[4744]: I1008 10:26:01.877813 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2821f947-c8b3-429f-995c-89dca9197d07-kube-api-access-q7r97" (OuterVolumeSpecName: "kube-api-access-q7r97") pod "2821f947-c8b3-429f-995c-89dca9197d07" (UID: "2821f947-c8b3-429f-995c-89dca9197d07"). InnerVolumeSpecName "kube-api-access-q7r97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:26:01 crc kubenswrapper[4744]: I1008 10:26:01.919508 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2821f947-c8b3-429f-995c-89dca9197d07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2821f947-c8b3-429f-995c-89dca9197d07" (UID: "2821f947-c8b3-429f-995c-89dca9197d07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:26:01 crc kubenswrapper[4744]: I1008 10:26:01.972898 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2821f947-c8b3-429f-995c-89dca9197d07-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:26:01 crc kubenswrapper[4744]: I1008 10:26:01.972947 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7r97\" (UniqueName: \"kubernetes.io/projected/2821f947-c8b3-429f-995c-89dca9197d07-kube-api-access-q7r97\") on node \"crc\" DevicePath \"\"" Oct 08 10:26:01 crc kubenswrapper[4744]: I1008 10:26:01.972961 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2821f947-c8b3-429f-995c-89dca9197d07-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.269351 4744 generic.go:334] "Generic (PLEG): container finished" podID="2821f947-c8b3-429f-995c-89dca9197d07" containerID="ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36" exitCode=0 Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.269418 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r966p" event={"ID":"2821f947-c8b3-429f-995c-89dca9197d07","Type":"ContainerDied","Data":"ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36"} Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.269487 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r966p" event={"ID":"2821f947-c8b3-429f-995c-89dca9197d07","Type":"ContainerDied","Data":"db053a2bf4570537b5d815897928c26e44ae5c236a9731a74a29643fca1d58aa"} Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.269506 4744 scope.go:117] "RemoveContainer" containerID="ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36" Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.269407 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r966p" Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.306220 4744 scope.go:117] "RemoveContainer" containerID="0fb326ea64e2e8f005a3805d8b90b997a09f46484354c3d54dbd49097f5580f2" Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.356478 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r966p"] Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.367529 4744 scope.go:117] "RemoveContainer" containerID="bf0ac8a1833dc99412c98f7cea16b85f42b80ebf156f3d698b81a3de2625160a" Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.379032 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r966p"] Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.411290 4744 scope.go:117] "RemoveContainer" containerID="ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36" Oct 08 10:26:02 crc kubenswrapper[4744]: E1008 10:26:02.412794 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36\": container with ID starting with ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36 not found: ID does not exist" containerID="ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36" Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.412835 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36"} err="failed to get container status \"ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36\": rpc error: code = NotFound desc = could not find container \"ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36\": container with ID starting with ee9941ed6022da76fc2a445910a30ca9d3e4ae013dd3b3fbbc9664946330de36 not found: ID does not exist" Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.412861 4744 scope.go:117] "RemoveContainer" containerID="0fb326ea64e2e8f005a3805d8b90b997a09f46484354c3d54dbd49097f5580f2" Oct 08 10:26:02 crc kubenswrapper[4744]: E1008 10:26:02.415500 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fb326ea64e2e8f005a3805d8b90b997a09f46484354c3d54dbd49097f5580f2\": container with ID starting with 0fb326ea64e2e8f005a3805d8b90b997a09f46484354c3d54dbd49097f5580f2 not found: ID does not exist" containerID="0fb326ea64e2e8f005a3805d8b90b997a09f46484354c3d54dbd49097f5580f2" Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.415538 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fb326ea64e2e8f005a3805d8b90b997a09f46484354c3d54dbd49097f5580f2"} err="failed to get container status \"0fb326ea64e2e8f005a3805d8b90b997a09f46484354c3d54dbd49097f5580f2\": rpc error: code = NotFound desc = could not find container \"0fb326ea64e2e8f005a3805d8b90b997a09f46484354c3d54dbd49097f5580f2\": container with ID starting with 0fb326ea64e2e8f005a3805d8b90b997a09f46484354c3d54dbd49097f5580f2 not found: ID does not exist" Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.415562 4744 scope.go:117] "RemoveContainer" containerID="bf0ac8a1833dc99412c98f7cea16b85f42b80ebf156f3d698b81a3de2625160a" Oct 08 10:26:02 crc kubenswrapper[4744]: E1008 10:26:02.416130 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf0ac8a1833dc99412c98f7cea16b85f42b80ebf156f3d698b81a3de2625160a\": container with ID starting with bf0ac8a1833dc99412c98f7cea16b85f42b80ebf156f3d698b81a3de2625160a not found: ID does not exist" containerID="bf0ac8a1833dc99412c98f7cea16b85f42b80ebf156f3d698b81a3de2625160a" Oct 08 10:26:02 crc kubenswrapper[4744]: I1008 10:26:02.416192 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf0ac8a1833dc99412c98f7cea16b85f42b80ebf156f3d698b81a3de2625160a"} err="failed to get container status \"bf0ac8a1833dc99412c98f7cea16b85f42b80ebf156f3d698b81a3de2625160a\": rpc error: code = NotFound desc = could not find container \"bf0ac8a1833dc99412c98f7cea16b85f42b80ebf156f3d698b81a3de2625160a\": container with ID starting with bf0ac8a1833dc99412c98f7cea16b85f42b80ebf156f3d698b81a3de2625160a not found: ID does not exist" Oct 08 10:26:03 crc kubenswrapper[4744]: I1008 10:26:03.478147 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2821f947-c8b3-429f-995c-89dca9197d07" path="/var/lib/kubelet/pods/2821f947-c8b3-429f-995c-89dca9197d07/volumes" Oct 08 10:26:05 crc kubenswrapper[4744]: I1008 10:26:05.464875 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:26:05 crc kubenswrapper[4744]: E1008 10:26:05.465686 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:26:14 crc kubenswrapper[4744]: I1008 10:26:14.049995 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-8rkml_4f0fc93a-5521-4a7a-90d7-7b1794dc728f/controller/0.log" Oct 08 10:26:14 crc kubenswrapper[4744]: I1008 10:26:14.090444 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-8rkml_4f0fc93a-5521-4a7a-90d7-7b1794dc728f/kube-rbac-proxy/0.log" Oct 08 10:26:14 crc kubenswrapper[4744]: I1008 10:26:14.273487 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-frr-files/0.log" Oct 08 10:26:14 crc kubenswrapper[4744]: I1008 10:26:14.743158 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-reloader/0.log" Oct 08 10:26:14 crc kubenswrapper[4744]: I1008 10:26:14.785341 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-metrics/0.log" Oct 08 10:26:14 crc kubenswrapper[4744]: I1008 10:26:14.808241 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-frr-files/0.log" Oct 08 10:26:14 crc kubenswrapper[4744]: I1008 10:26:14.834093 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-reloader/0.log" Oct 08 10:26:15 crc kubenswrapper[4744]: I1008 10:26:15.094237 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-frr-files/0.log" Oct 08 10:26:15 crc kubenswrapper[4744]: I1008 10:26:15.126236 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-reloader/0.log" Oct 08 10:26:15 crc kubenswrapper[4744]: I1008 10:26:15.182647 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-metrics/0.log" Oct 08 10:26:15 crc kubenswrapper[4744]: I1008 10:26:15.184643 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-metrics/0.log" Oct 08 10:26:15 crc kubenswrapper[4744]: I1008 10:26:15.370963 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-reloader/0.log" Oct 08 10:26:15 crc kubenswrapper[4744]: I1008 10:26:15.392793 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/controller/0.log" Oct 08 10:26:15 crc kubenswrapper[4744]: I1008 10:26:15.424583 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-metrics/0.log" Oct 08 10:26:15 crc kubenswrapper[4744]: I1008 10:26:15.438474 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-frr-files/0.log" Oct 08 10:26:15 crc kubenswrapper[4744]: I1008 10:26:15.645009 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/kube-rbac-proxy/0.log" Oct 08 10:26:15 crc kubenswrapper[4744]: I1008 10:26:15.701621 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/kube-rbac-proxy-frr/0.log" Oct 08 10:26:15 crc kubenswrapper[4744]: I1008 10:26:15.713250 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/frr-metrics/0.log" Oct 08 10:26:16 crc kubenswrapper[4744]: I1008 10:26:16.011821 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/reloader/0.log" Oct 08 10:26:16 crc kubenswrapper[4744]: I1008 10:26:16.097585 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-ptt9h_cf7d04b2-1136-4427-8939-e2974337725c/frr-k8s-webhook-server/0.log" Oct 08 10:26:16 crc kubenswrapper[4744]: I1008 10:26:16.285556 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5998774b9-gk4vj_fb79b0ea-df0c-4f82-a297-4ce347bdc363/manager/0.log" Oct 08 10:26:16 crc kubenswrapper[4744]: I1008 10:26:16.677706 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-dbd787fb9-lsv2s_9de92423-8f60-43bd-842c-b47f51c9e20c/webhook-server/0.log" Oct 08 10:26:16 crc kubenswrapper[4744]: I1008 10:26:16.895101 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-89qxn_349b3e3f-87d0-4e3d-b2dc-c96452a7199c/kube-rbac-proxy/0.log" Oct 08 10:26:17 crc kubenswrapper[4744]: I1008 10:26:17.055936 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/frr/0.log" Oct 08 10:26:17 crc kubenswrapper[4744]: I1008 10:26:17.312533 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-89qxn_349b3e3f-87d0-4e3d-b2dc-c96452a7199c/speaker/0.log" Oct 08 10:26:17 crc kubenswrapper[4744]: I1008 10:26:17.453749 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:26:17 crc kubenswrapper[4744]: E1008 10:26:17.454007 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:26:31 crc kubenswrapper[4744]: I1008 10:26:31.234316 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/util/0.log" Oct 08 10:26:31 crc kubenswrapper[4744]: I1008 10:26:31.499180 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/util/0.log" Oct 08 10:26:31 crc kubenswrapper[4744]: I1008 10:26:31.526168 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/pull/0.log" Oct 08 10:26:31 crc kubenswrapper[4744]: I1008 10:26:31.588954 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/pull/0.log" Oct 08 10:26:31 crc kubenswrapper[4744]: I1008 10:26:31.743881 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/util/0.log" Oct 08 10:26:31 crc kubenswrapper[4744]: I1008 10:26:31.761865 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/extract/0.log" Oct 08 10:26:31 crc kubenswrapper[4744]: I1008 10:26:31.827620 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/pull/0.log" Oct 08 10:26:31 crc kubenswrapper[4744]: I1008 10:26:31.982246 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/extract-utilities/0.log" Oct 08 10:26:32 crc kubenswrapper[4744]: I1008 10:26:32.217619 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/extract-utilities/0.log" Oct 08 10:26:32 crc kubenswrapper[4744]: I1008 10:26:32.247654 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/extract-content/0.log" Oct 08 10:26:32 crc kubenswrapper[4744]: I1008 10:26:32.269897 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/extract-content/0.log" Oct 08 10:26:32 crc kubenswrapper[4744]: I1008 10:26:32.428425 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/extract-utilities/0.log" Oct 08 10:26:32 crc kubenswrapper[4744]: I1008 10:26:32.453531 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/extract-content/0.log" Oct 08 10:26:32 crc kubenswrapper[4744]: I1008 10:26:32.454390 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:26:32 crc kubenswrapper[4744]: E1008 10:26:32.454636 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:26:32 crc kubenswrapper[4744]: I1008 10:26:32.850670 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/extract-utilities/0.log" Oct 08 10:26:33 crc kubenswrapper[4744]: I1008 10:26:33.018463 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/extract-content/0.log" Oct 08 10:26:33 crc kubenswrapper[4744]: I1008 10:26:33.048532 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/extract-utilities/0.log" Oct 08 10:26:33 crc kubenswrapper[4744]: I1008 10:26:33.120141 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/registry-server/0.log" Oct 08 10:26:33 crc kubenswrapper[4744]: I1008 10:26:33.190349 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/extract-content/0.log" Oct 08 10:26:33 crc kubenswrapper[4744]: I1008 10:26:33.397206 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/extract-utilities/0.log" Oct 08 10:26:33 crc kubenswrapper[4744]: I1008 10:26:33.448179 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/extract-content/0.log" Oct 08 10:26:33 crc kubenswrapper[4744]: I1008 10:26:33.946151 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/util/0.log" Oct 08 10:26:34 crc kubenswrapper[4744]: I1008 10:26:34.000222 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/registry-server/0.log" Oct 08 10:26:34 crc kubenswrapper[4744]: I1008 10:26:34.044604 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/pull/0.log" Oct 08 10:26:34 crc kubenswrapper[4744]: I1008 10:26:34.109554 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/util/0.log" Oct 08 10:26:34 crc kubenswrapper[4744]: I1008 10:26:34.205026 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/pull/0.log" Oct 08 10:26:34 crc kubenswrapper[4744]: I1008 10:26:34.385081 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/extract/0.log" Oct 08 10:26:34 crc kubenswrapper[4744]: I1008 10:26:34.403604 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/util/0.log" Oct 08 10:26:34 crc kubenswrapper[4744]: I1008 10:26:34.460450 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/pull/0.log" Oct 08 10:26:34 crc kubenswrapper[4744]: I1008 10:26:34.608642 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-rjzwz_0364f287-922c-410f-9cee-1866cdba42e1/marketplace-operator/0.log" Oct 08 10:26:34 crc kubenswrapper[4744]: I1008 10:26:34.667416 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/extract-utilities/0.log" Oct 08 10:26:34 crc kubenswrapper[4744]: I1008 10:26:34.876169 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/extract-content/0.log" Oct 08 10:26:34 crc kubenswrapper[4744]: I1008 10:26:34.894227 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/extract-content/0.log" Oct 08 10:26:34 crc kubenswrapper[4744]: I1008 10:26:34.894275 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/extract-utilities/0.log" Oct 08 10:26:35 crc kubenswrapper[4744]: I1008 10:26:35.157671 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/extract-content/0.log" Oct 08 10:26:35 crc kubenswrapper[4744]: I1008 10:26:35.187245 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/extract-utilities/0.log" Oct 08 10:26:35 crc kubenswrapper[4744]: I1008 10:26:35.378424 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/extract-utilities/0.log" Oct 08 10:26:35 crc kubenswrapper[4744]: I1008 10:26:35.403243 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/registry-server/0.log" Oct 08 10:26:35 crc kubenswrapper[4744]: I1008 10:26:35.967329 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/extract-content/0.log" Oct 08 10:26:35 crc kubenswrapper[4744]: I1008 10:26:35.985022 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/extract-content/0.log" Oct 08 10:26:36 crc kubenswrapper[4744]: I1008 10:26:36.000635 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/extract-utilities/0.log" Oct 08 10:26:36 crc kubenswrapper[4744]: I1008 10:26:36.154261 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/extract-utilities/0.log" Oct 08 10:26:36 crc kubenswrapper[4744]: I1008 10:26:36.176604 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/extract-content/0.log" Oct 08 10:26:36 crc kubenswrapper[4744]: I1008 10:26:36.586057 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/registry-server/0.log" Oct 08 10:26:44 crc kubenswrapper[4744]: I1008 10:26:44.453785 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:26:44 crc kubenswrapper[4744]: E1008 10:26:44.455043 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:26:56 crc kubenswrapper[4744]: I1008 10:26:56.453492 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:26:56 crc kubenswrapper[4744]: E1008 10:26:56.454342 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:27:11 crc kubenswrapper[4744]: I1008 10:27:11.454058 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:27:11 crc kubenswrapper[4744]: E1008 10:27:11.454854 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:27:26 crc kubenswrapper[4744]: I1008 10:27:26.454482 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:27:26 crc kubenswrapper[4744]: E1008 10:27:26.455146 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:27:37 crc kubenswrapper[4744]: I1008 10:27:37.452760 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:27:37 crc kubenswrapper[4744]: E1008 10:27:37.453802 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:27:49 crc kubenswrapper[4744]: I1008 10:27:49.455994 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:27:49 crc kubenswrapper[4744]: E1008 10:27:49.456709 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:28:02 crc kubenswrapper[4744]: I1008 10:28:02.452936 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:28:02 crc kubenswrapper[4744]: E1008 10:28:02.453807 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:28:16 crc kubenswrapper[4744]: I1008 10:28:16.454950 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:28:16 crc kubenswrapper[4744]: E1008 10:28:16.455454 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.593307 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bf7k9"] Oct 08 10:28:29 crc kubenswrapper[4744]: E1008 10:28:29.594370 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2af19560-76bd-4524-b537-ac61cfe6730a" containerName="extract-content" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.594476 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2af19560-76bd-4524-b537-ac61cfe6730a" containerName="extract-content" Oct 08 10:28:29 crc kubenswrapper[4744]: E1008 10:28:29.594497 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2af19560-76bd-4524-b537-ac61cfe6730a" containerName="registry-server" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.594508 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2af19560-76bd-4524-b537-ac61cfe6730a" containerName="registry-server" Oct 08 10:28:29 crc kubenswrapper[4744]: E1008 10:28:29.594524 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2821f947-c8b3-429f-995c-89dca9197d07" containerName="extract-utilities" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.594532 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2821f947-c8b3-429f-995c-89dca9197d07" containerName="extract-utilities" Oct 08 10:28:29 crc kubenswrapper[4744]: E1008 10:28:29.594577 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2821f947-c8b3-429f-995c-89dca9197d07" containerName="extract-content" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.594587 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2821f947-c8b3-429f-995c-89dca9197d07" containerName="extract-content" Oct 08 10:28:29 crc kubenswrapper[4744]: E1008 10:28:29.594601 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2821f947-c8b3-429f-995c-89dca9197d07" containerName="registry-server" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.594608 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2821f947-c8b3-429f-995c-89dca9197d07" containerName="registry-server" Oct 08 10:28:29 crc kubenswrapper[4744]: E1008 10:28:29.594629 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2af19560-76bd-4524-b537-ac61cfe6730a" containerName="extract-utilities" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.594637 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2af19560-76bd-4524-b537-ac61cfe6730a" containerName="extract-utilities" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.594885 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2af19560-76bd-4524-b537-ac61cfe6730a" containerName="registry-server" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.594909 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2821f947-c8b3-429f-995c-89dca9197d07" containerName="registry-server" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.602105 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bf7k9"] Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.602444 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.772666 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e35b90-83a9-4e38-85dd-147b0c933b4b-utilities\") pod \"redhat-marketplace-bf7k9\" (UID: \"83e35b90-83a9-4e38-85dd-147b0c933b4b\") " pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.772971 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e35b90-83a9-4e38-85dd-147b0c933b4b-catalog-content\") pod \"redhat-marketplace-bf7k9\" (UID: \"83e35b90-83a9-4e38-85dd-147b0c933b4b\") " pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.773083 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwnzg\" (UniqueName: \"kubernetes.io/projected/83e35b90-83a9-4e38-85dd-147b0c933b4b-kube-api-access-vwnzg\") pod \"redhat-marketplace-bf7k9\" (UID: \"83e35b90-83a9-4e38-85dd-147b0c933b4b\") " pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.874438 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e35b90-83a9-4e38-85dd-147b0c933b4b-utilities\") pod \"redhat-marketplace-bf7k9\" (UID: \"83e35b90-83a9-4e38-85dd-147b0c933b4b\") " pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.874487 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e35b90-83a9-4e38-85dd-147b0c933b4b-catalog-content\") pod \"redhat-marketplace-bf7k9\" (UID: \"83e35b90-83a9-4e38-85dd-147b0c933b4b\") " pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.874558 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwnzg\" (UniqueName: \"kubernetes.io/projected/83e35b90-83a9-4e38-85dd-147b0c933b4b-kube-api-access-vwnzg\") pod \"redhat-marketplace-bf7k9\" (UID: \"83e35b90-83a9-4e38-85dd-147b0c933b4b\") " pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.875284 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e35b90-83a9-4e38-85dd-147b0c933b4b-utilities\") pod \"redhat-marketplace-bf7k9\" (UID: \"83e35b90-83a9-4e38-85dd-147b0c933b4b\") " pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.875315 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e35b90-83a9-4e38-85dd-147b0c933b4b-catalog-content\") pod \"redhat-marketplace-bf7k9\" (UID: \"83e35b90-83a9-4e38-85dd-147b0c933b4b\") " pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.896465 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwnzg\" (UniqueName: \"kubernetes.io/projected/83e35b90-83a9-4e38-85dd-147b0c933b4b-kube-api-access-vwnzg\") pod \"redhat-marketplace-bf7k9\" (UID: \"83e35b90-83a9-4e38-85dd-147b0c933b4b\") " pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:29 crc kubenswrapper[4744]: I1008 10:28:29.935018 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:30 crc kubenswrapper[4744]: I1008 10:28:30.458194 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bf7k9"] Oct 08 10:28:30 crc kubenswrapper[4744]: W1008 10:28:30.471996 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83e35b90_83a9_4e38_85dd_147b0c933b4b.slice/crio-8646f18f5d28afe8a5ecf17b1b1fef689d129b58956b924c380a5e8fee435f7b WatchSource:0}: Error finding container 8646f18f5d28afe8a5ecf17b1b1fef689d129b58956b924c380a5e8fee435f7b: Status 404 returned error can't find the container with id 8646f18f5d28afe8a5ecf17b1b1fef689d129b58956b924c380a5e8fee435f7b Oct 08 10:28:30 crc kubenswrapper[4744]: I1008 10:28:30.647805 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf7k9" event={"ID":"83e35b90-83a9-4e38-85dd-147b0c933b4b","Type":"ContainerStarted","Data":"8646f18f5d28afe8a5ecf17b1b1fef689d129b58956b924c380a5e8fee435f7b"} Oct 08 10:28:31 crc kubenswrapper[4744]: I1008 10:28:31.453329 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:28:31 crc kubenswrapper[4744]: E1008 10:28:31.453612 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:28:31 crc kubenswrapper[4744]: I1008 10:28:31.656804 4744 generic.go:334] "Generic (PLEG): container finished" podID="83e35b90-83a9-4e38-85dd-147b0c933b4b" containerID="c79ba23e02998a0e6daac1e307ff112b1b824a304abebaf0b3a90530eae3a7b0" exitCode=0 Oct 08 10:28:31 crc kubenswrapper[4744]: I1008 10:28:31.656861 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf7k9" event={"ID":"83e35b90-83a9-4e38-85dd-147b0c933b4b","Type":"ContainerDied","Data":"c79ba23e02998a0e6daac1e307ff112b1b824a304abebaf0b3a90530eae3a7b0"} Oct 08 10:28:31 crc kubenswrapper[4744]: I1008 10:28:31.660112 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 10:28:33 crc kubenswrapper[4744]: I1008 10:28:33.678147 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf7k9" event={"ID":"83e35b90-83a9-4e38-85dd-147b0c933b4b","Type":"ContainerStarted","Data":"9aed4ee076791f08162f961189278af4ba396a09743d2c1fd545832cabe2e12f"} Oct 08 10:28:34 crc kubenswrapper[4744]: I1008 10:28:34.687951 4744 generic.go:334] "Generic (PLEG): container finished" podID="83e35b90-83a9-4e38-85dd-147b0c933b4b" containerID="9aed4ee076791f08162f961189278af4ba396a09743d2c1fd545832cabe2e12f" exitCode=0 Oct 08 10:28:34 crc kubenswrapper[4744]: I1008 10:28:34.688005 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf7k9" event={"ID":"83e35b90-83a9-4e38-85dd-147b0c933b4b","Type":"ContainerDied","Data":"9aed4ee076791f08162f961189278af4ba396a09743d2c1fd545832cabe2e12f"} Oct 08 10:28:36 crc kubenswrapper[4744]: I1008 10:28:36.705203 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf7k9" event={"ID":"83e35b90-83a9-4e38-85dd-147b0c933b4b","Type":"ContainerStarted","Data":"ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960"} Oct 08 10:28:36 crc kubenswrapper[4744]: I1008 10:28:36.727495 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bf7k9" podStartSLOduration=3.314835127 podStartE2EDuration="7.727476246s" podCreationTimestamp="2025-10-08 10:28:29 +0000 UTC" firstStartedPulling="2025-10-08 10:28:31.658586823 +0000 UTC m=+4606.906232062" lastFinishedPulling="2025-10-08 10:28:36.071227932 +0000 UTC m=+4611.318873181" observedRunningTime="2025-10-08 10:28:36.726838578 +0000 UTC m=+4611.974483817" watchObservedRunningTime="2025-10-08 10:28:36.727476246 +0000 UTC m=+4611.975121485" Oct 08 10:28:39 crc kubenswrapper[4744]: I1008 10:28:39.935820 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:39 crc kubenswrapper[4744]: I1008 10:28:39.936501 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:39 crc kubenswrapper[4744]: I1008 10:28:39.993598 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:46 crc kubenswrapper[4744]: I1008 10:28:46.453335 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:28:46 crc kubenswrapper[4744]: E1008 10:28:46.454162 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:28:50 crc kubenswrapper[4744]: I1008 10:28:50.000563 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:50 crc kubenswrapper[4744]: I1008 10:28:50.074344 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bf7k9"] Oct 08 10:28:50 crc kubenswrapper[4744]: I1008 10:28:50.863635 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bf7k9" podUID="83e35b90-83a9-4e38-85dd-147b0c933b4b" containerName="registry-server" containerID="cri-o://ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960" gracePeriod=2 Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.373684 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.474355 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e35b90-83a9-4e38-85dd-147b0c933b4b-catalog-content\") pod \"83e35b90-83a9-4e38-85dd-147b0c933b4b\" (UID: \"83e35b90-83a9-4e38-85dd-147b0c933b4b\") " Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.489554 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83e35b90-83a9-4e38-85dd-147b0c933b4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83e35b90-83a9-4e38-85dd-147b0c933b4b" (UID: "83e35b90-83a9-4e38-85dd-147b0c933b4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.575659 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwnzg\" (UniqueName: \"kubernetes.io/projected/83e35b90-83a9-4e38-85dd-147b0c933b4b-kube-api-access-vwnzg\") pod \"83e35b90-83a9-4e38-85dd-147b0c933b4b\" (UID: \"83e35b90-83a9-4e38-85dd-147b0c933b4b\") " Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.575769 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e35b90-83a9-4e38-85dd-147b0c933b4b-utilities\") pod \"83e35b90-83a9-4e38-85dd-147b0c933b4b\" (UID: \"83e35b90-83a9-4e38-85dd-147b0c933b4b\") " Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.576272 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e35b90-83a9-4e38-85dd-147b0c933b4b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.576526 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83e35b90-83a9-4e38-85dd-147b0c933b4b-utilities" (OuterVolumeSpecName: "utilities") pod "83e35b90-83a9-4e38-85dd-147b0c933b4b" (UID: "83e35b90-83a9-4e38-85dd-147b0c933b4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.589100 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83e35b90-83a9-4e38-85dd-147b0c933b4b-kube-api-access-vwnzg" (OuterVolumeSpecName: "kube-api-access-vwnzg") pod "83e35b90-83a9-4e38-85dd-147b0c933b4b" (UID: "83e35b90-83a9-4e38-85dd-147b0c933b4b"). InnerVolumeSpecName "kube-api-access-vwnzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.678435 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwnzg\" (UniqueName: \"kubernetes.io/projected/83e35b90-83a9-4e38-85dd-147b0c933b4b-kube-api-access-vwnzg\") on node \"crc\" DevicePath \"\"" Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.678470 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e35b90-83a9-4e38-85dd-147b0c933b4b-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.878789 4744 generic.go:334] "Generic (PLEG): container finished" podID="83e35b90-83a9-4e38-85dd-147b0c933b4b" containerID="ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960" exitCode=0 Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.879141 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf7k9" event={"ID":"83e35b90-83a9-4e38-85dd-147b0c933b4b","Type":"ContainerDied","Data":"ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960"} Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.879185 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bf7k9" event={"ID":"83e35b90-83a9-4e38-85dd-147b0c933b4b","Type":"ContainerDied","Data":"8646f18f5d28afe8a5ecf17b1b1fef689d129b58956b924c380a5e8fee435f7b"} Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.879202 4744 scope.go:117] "RemoveContainer" containerID="ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960" Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.879358 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bf7k9" Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.932138 4744 scope.go:117] "RemoveContainer" containerID="9aed4ee076791f08162f961189278af4ba396a09743d2c1fd545832cabe2e12f" Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.934544 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bf7k9"] Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.945938 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bf7k9"] Oct 08 10:28:51 crc kubenswrapper[4744]: I1008 10:28:51.969784 4744 scope.go:117] "RemoveContainer" containerID="c79ba23e02998a0e6daac1e307ff112b1b824a304abebaf0b3a90530eae3a7b0" Oct 08 10:28:52 crc kubenswrapper[4744]: I1008 10:28:52.002045 4744 scope.go:117] "RemoveContainer" containerID="ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960" Oct 08 10:28:52 crc kubenswrapper[4744]: E1008 10:28:52.002778 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960\": container with ID starting with ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960 not found: ID does not exist" containerID="ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960" Oct 08 10:28:52 crc kubenswrapper[4744]: I1008 10:28:52.002820 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960"} err="failed to get container status \"ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960\": rpc error: code = NotFound desc = could not find container \"ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960\": container with ID starting with ecb23488847737dad400cfa2989d6cbe8b245bdff96bfe2c7d1fc4894f344960 not found: ID does not exist" Oct 08 10:28:52 crc kubenswrapper[4744]: I1008 10:28:52.002849 4744 scope.go:117] "RemoveContainer" containerID="9aed4ee076791f08162f961189278af4ba396a09743d2c1fd545832cabe2e12f" Oct 08 10:28:52 crc kubenswrapper[4744]: E1008 10:28:52.003173 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aed4ee076791f08162f961189278af4ba396a09743d2c1fd545832cabe2e12f\": container with ID starting with 9aed4ee076791f08162f961189278af4ba396a09743d2c1fd545832cabe2e12f not found: ID does not exist" containerID="9aed4ee076791f08162f961189278af4ba396a09743d2c1fd545832cabe2e12f" Oct 08 10:28:52 crc kubenswrapper[4744]: I1008 10:28:52.003234 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aed4ee076791f08162f961189278af4ba396a09743d2c1fd545832cabe2e12f"} err="failed to get container status \"9aed4ee076791f08162f961189278af4ba396a09743d2c1fd545832cabe2e12f\": rpc error: code = NotFound desc = could not find container \"9aed4ee076791f08162f961189278af4ba396a09743d2c1fd545832cabe2e12f\": container with ID starting with 9aed4ee076791f08162f961189278af4ba396a09743d2c1fd545832cabe2e12f not found: ID does not exist" Oct 08 10:28:52 crc kubenswrapper[4744]: I1008 10:28:52.003255 4744 scope.go:117] "RemoveContainer" containerID="c79ba23e02998a0e6daac1e307ff112b1b824a304abebaf0b3a90530eae3a7b0" Oct 08 10:28:52 crc kubenswrapper[4744]: E1008 10:28:52.003992 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c79ba23e02998a0e6daac1e307ff112b1b824a304abebaf0b3a90530eae3a7b0\": container with ID starting with c79ba23e02998a0e6daac1e307ff112b1b824a304abebaf0b3a90530eae3a7b0 not found: ID does not exist" containerID="c79ba23e02998a0e6daac1e307ff112b1b824a304abebaf0b3a90530eae3a7b0" Oct 08 10:28:52 crc kubenswrapper[4744]: I1008 10:28:52.004053 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79ba23e02998a0e6daac1e307ff112b1b824a304abebaf0b3a90530eae3a7b0"} err="failed to get container status \"c79ba23e02998a0e6daac1e307ff112b1b824a304abebaf0b3a90530eae3a7b0\": rpc error: code = NotFound desc = could not find container \"c79ba23e02998a0e6daac1e307ff112b1b824a304abebaf0b3a90530eae3a7b0\": container with ID starting with c79ba23e02998a0e6daac1e307ff112b1b824a304abebaf0b3a90530eae3a7b0 not found: ID does not exist" Oct 08 10:28:53 crc kubenswrapper[4744]: I1008 10:28:53.467973 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83e35b90-83a9-4e38-85dd-147b0c933b4b" path="/var/lib/kubelet/pods/83e35b90-83a9-4e38-85dd-147b0c933b4b/volumes" Oct 08 10:28:58 crc kubenswrapper[4744]: I1008 10:28:58.453205 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:28:58 crc kubenswrapper[4744]: E1008 10:28:58.454128 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:28:59 crc kubenswrapper[4744]: I1008 10:28:59.355598 4744 scope.go:117] "RemoveContainer" containerID="2ce2e5bed34ee83a1fba64376cde9515a540d0409a38130a83dde8ee3d95eb0c" Oct 08 10:29:10 crc kubenswrapper[4744]: I1008 10:29:10.454656 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:29:10 crc kubenswrapper[4744]: E1008 10:29:10.455971 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:29:22 crc kubenswrapper[4744]: I1008 10:29:22.455081 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:29:22 crc kubenswrapper[4744]: E1008 10:29:22.455854 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:29:25 crc kubenswrapper[4744]: I1008 10:29:25.200779 4744 generic.go:334] "Generic (PLEG): container finished" podID="a365de61-417b-4735-b90a-e8a8feb93085" containerID="f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1" exitCode=0 Oct 08 10:29:25 crc kubenswrapper[4744]: I1008 10:29:25.200858 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lx88m/must-gather-wrnk6" event={"ID":"a365de61-417b-4735-b90a-e8a8feb93085","Type":"ContainerDied","Data":"f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1"} Oct 08 10:29:25 crc kubenswrapper[4744]: I1008 10:29:25.201802 4744 scope.go:117] "RemoveContainer" containerID="f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1" Oct 08 10:29:25 crc kubenswrapper[4744]: I1008 10:29:25.775760 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lx88m_must-gather-wrnk6_a365de61-417b-4735-b90a-e8a8feb93085/gather/0.log" Oct 08 10:29:34 crc kubenswrapper[4744]: I1008 10:29:34.530924 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lx88m/must-gather-wrnk6"] Oct 08 10:29:34 crc kubenswrapper[4744]: I1008 10:29:34.532725 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-lx88m/must-gather-wrnk6" podUID="a365de61-417b-4735-b90a-e8a8feb93085" containerName="copy" containerID="cri-o://a19df090fa08ea0ecad3c06c82d8f110bc82ab2e7b14404d061e3b2acab79a9f" gracePeriod=2 Oct 08 10:29:34 crc kubenswrapper[4744]: I1008 10:29:34.538324 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lx88m/must-gather-wrnk6"] Oct 08 10:29:34 crc kubenswrapper[4744]: I1008 10:29:34.970338 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lx88m_must-gather-wrnk6_a365de61-417b-4735-b90a-e8a8feb93085/copy/0.log" Oct 08 10:29:34 crc kubenswrapper[4744]: I1008 10:29:34.971046 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/must-gather-wrnk6" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.099585 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a365de61-417b-4735-b90a-e8a8feb93085-must-gather-output\") pod \"a365de61-417b-4735-b90a-e8a8feb93085\" (UID: \"a365de61-417b-4735-b90a-e8a8feb93085\") " Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.099860 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9vn4\" (UniqueName: \"kubernetes.io/projected/a365de61-417b-4735-b90a-e8a8feb93085-kube-api-access-r9vn4\") pod \"a365de61-417b-4735-b90a-e8a8feb93085\" (UID: \"a365de61-417b-4735-b90a-e8a8feb93085\") " Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.106890 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a365de61-417b-4735-b90a-e8a8feb93085-kube-api-access-r9vn4" (OuterVolumeSpecName: "kube-api-access-r9vn4") pod "a365de61-417b-4735-b90a-e8a8feb93085" (UID: "a365de61-417b-4735-b90a-e8a8feb93085"). InnerVolumeSpecName "kube-api-access-r9vn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.201901 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9vn4\" (UniqueName: \"kubernetes.io/projected/a365de61-417b-4735-b90a-e8a8feb93085-kube-api-access-r9vn4\") on node \"crc\" DevicePath \"\"" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.290637 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a365de61-417b-4735-b90a-e8a8feb93085-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a365de61-417b-4735-b90a-e8a8feb93085" (UID: "a365de61-417b-4735-b90a-e8a8feb93085"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.297246 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lx88m_must-gather-wrnk6_a365de61-417b-4735-b90a-e8a8feb93085/copy/0.log" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.297693 4744 generic.go:334] "Generic (PLEG): container finished" podID="a365de61-417b-4735-b90a-e8a8feb93085" containerID="a19df090fa08ea0ecad3c06c82d8f110bc82ab2e7b14404d061e3b2acab79a9f" exitCode=143 Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.297742 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lx88m/must-gather-wrnk6" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.297747 4744 scope.go:117] "RemoveContainer" containerID="a19df090fa08ea0ecad3c06c82d8f110bc82ab2e7b14404d061e3b2acab79a9f" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.304337 4744 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a365de61-417b-4735-b90a-e8a8feb93085-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.317613 4744 scope.go:117] "RemoveContainer" containerID="f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.355536 4744 scope.go:117] "RemoveContainer" containerID="a19df090fa08ea0ecad3c06c82d8f110bc82ab2e7b14404d061e3b2acab79a9f" Oct 08 10:29:35 crc kubenswrapper[4744]: E1008 10:29:35.356137 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a19df090fa08ea0ecad3c06c82d8f110bc82ab2e7b14404d061e3b2acab79a9f\": container with ID starting with a19df090fa08ea0ecad3c06c82d8f110bc82ab2e7b14404d061e3b2acab79a9f not found: ID does not exist" containerID="a19df090fa08ea0ecad3c06c82d8f110bc82ab2e7b14404d061e3b2acab79a9f" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.356187 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a19df090fa08ea0ecad3c06c82d8f110bc82ab2e7b14404d061e3b2acab79a9f"} err="failed to get container status \"a19df090fa08ea0ecad3c06c82d8f110bc82ab2e7b14404d061e3b2acab79a9f\": rpc error: code = NotFound desc = could not find container \"a19df090fa08ea0ecad3c06c82d8f110bc82ab2e7b14404d061e3b2acab79a9f\": container with ID starting with a19df090fa08ea0ecad3c06c82d8f110bc82ab2e7b14404d061e3b2acab79a9f not found: ID does not exist" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.356215 4744 scope.go:117] "RemoveContainer" containerID="f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1" Oct 08 10:29:35 crc kubenswrapper[4744]: E1008 10:29:35.356719 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1\": container with ID starting with f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1 not found: ID does not exist" containerID="f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.356770 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1"} err="failed to get container status \"f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1\": rpc error: code = NotFound desc = could not find container \"f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1\": container with ID starting with f9e3ebc98a79d9c03f3fdf66a2e6fde3724474d16c29e73b38bbddf820ce5dd1 not found: ID does not exist" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.459445 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:29:35 crc kubenswrapper[4744]: E1008 10:29:35.459993 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:29:35 crc kubenswrapper[4744]: I1008 10:29:35.463164 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a365de61-417b-4735-b90a-e8a8feb93085" path="/var/lib/kubelet/pods/a365de61-417b-4735-b90a-e8a8feb93085/volumes" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.947106 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jzt92"] Oct 08 10:29:43 crc kubenswrapper[4744]: E1008 10:29:43.947980 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a365de61-417b-4735-b90a-e8a8feb93085" containerName="gather" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.947992 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a365de61-417b-4735-b90a-e8a8feb93085" containerName="gather" Oct 08 10:29:43 crc kubenswrapper[4744]: E1008 10:29:43.948002 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e35b90-83a9-4e38-85dd-147b0c933b4b" containerName="registry-server" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.948008 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e35b90-83a9-4e38-85dd-147b0c933b4b" containerName="registry-server" Oct 08 10:29:43 crc kubenswrapper[4744]: E1008 10:29:43.948017 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e35b90-83a9-4e38-85dd-147b0c933b4b" containerName="extract-utilities" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.948024 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e35b90-83a9-4e38-85dd-147b0c933b4b" containerName="extract-utilities" Oct 08 10:29:43 crc kubenswrapper[4744]: E1008 10:29:43.948057 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a365de61-417b-4735-b90a-e8a8feb93085" containerName="copy" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.948063 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a365de61-417b-4735-b90a-e8a8feb93085" containerName="copy" Oct 08 10:29:43 crc kubenswrapper[4744]: E1008 10:29:43.948073 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e35b90-83a9-4e38-85dd-147b0c933b4b" containerName="extract-content" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.948080 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e35b90-83a9-4e38-85dd-147b0c933b4b" containerName="extract-content" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.948269 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a365de61-417b-4735-b90a-e8a8feb93085" containerName="gather" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.948280 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a365de61-417b-4735-b90a-e8a8feb93085" containerName="copy" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.948295 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="83e35b90-83a9-4e38-85dd-147b0c933b4b" containerName="registry-server" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.949618 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.964638 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jzt92"] Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.974938 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3972149-5648-4a77-ae04-1153a8a32565-catalog-content\") pod \"redhat-operators-jzt92\" (UID: \"e3972149-5648-4a77-ae04-1153a8a32565\") " pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.975085 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvxct\" (UniqueName: \"kubernetes.io/projected/e3972149-5648-4a77-ae04-1153a8a32565-kube-api-access-nvxct\") pod \"redhat-operators-jzt92\" (UID: \"e3972149-5648-4a77-ae04-1153a8a32565\") " pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:29:43 crc kubenswrapper[4744]: I1008 10:29:43.975109 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3972149-5648-4a77-ae04-1153a8a32565-utilities\") pod \"redhat-operators-jzt92\" (UID: \"e3972149-5648-4a77-ae04-1153a8a32565\") " pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:29:44 crc kubenswrapper[4744]: I1008 10:29:44.077315 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvxct\" (UniqueName: \"kubernetes.io/projected/e3972149-5648-4a77-ae04-1153a8a32565-kube-api-access-nvxct\") pod \"redhat-operators-jzt92\" (UID: \"e3972149-5648-4a77-ae04-1153a8a32565\") " pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:29:44 crc kubenswrapper[4744]: I1008 10:29:44.077633 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3972149-5648-4a77-ae04-1153a8a32565-utilities\") pod \"redhat-operators-jzt92\" (UID: \"e3972149-5648-4a77-ae04-1153a8a32565\") " pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:29:44 crc kubenswrapper[4744]: I1008 10:29:44.077802 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3972149-5648-4a77-ae04-1153a8a32565-catalog-content\") pod \"redhat-operators-jzt92\" (UID: \"e3972149-5648-4a77-ae04-1153a8a32565\") " pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:29:44 crc kubenswrapper[4744]: I1008 10:29:44.078354 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3972149-5648-4a77-ae04-1153a8a32565-catalog-content\") pod \"redhat-operators-jzt92\" (UID: \"e3972149-5648-4a77-ae04-1153a8a32565\") " pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:29:44 crc kubenswrapper[4744]: I1008 10:29:44.078352 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3972149-5648-4a77-ae04-1153a8a32565-utilities\") pod \"redhat-operators-jzt92\" (UID: \"e3972149-5648-4a77-ae04-1153a8a32565\") " pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:29:44 crc kubenswrapper[4744]: I1008 10:29:44.099562 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvxct\" (UniqueName: \"kubernetes.io/projected/e3972149-5648-4a77-ae04-1153a8a32565-kube-api-access-nvxct\") pod \"redhat-operators-jzt92\" (UID: \"e3972149-5648-4a77-ae04-1153a8a32565\") " pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:29:44 crc kubenswrapper[4744]: I1008 10:29:44.291012 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:29:44 crc kubenswrapper[4744]: I1008 10:29:44.803077 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jzt92"] Oct 08 10:29:45 crc kubenswrapper[4744]: I1008 10:29:45.403069 4744 generic.go:334] "Generic (PLEG): container finished" podID="e3972149-5648-4a77-ae04-1153a8a32565" containerID="cf8dae41f4dc6c9dc737b675ccdc286868b47d710c677260d0e70a090d00f44e" exitCode=0 Oct 08 10:29:45 crc kubenswrapper[4744]: I1008 10:29:45.403220 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzt92" event={"ID":"e3972149-5648-4a77-ae04-1153a8a32565","Type":"ContainerDied","Data":"cf8dae41f4dc6c9dc737b675ccdc286868b47d710c677260d0e70a090d00f44e"} Oct 08 10:29:45 crc kubenswrapper[4744]: I1008 10:29:45.403350 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzt92" event={"ID":"e3972149-5648-4a77-ae04-1153a8a32565","Type":"ContainerStarted","Data":"5599f19ee345e7d96d14dc869c77ec209307db9fd975143d108bc7d9fad1f53d"} Oct 08 10:29:46 crc kubenswrapper[4744]: I1008 10:29:46.453479 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:29:46 crc kubenswrapper[4744]: E1008 10:29:46.453809 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:29:47 crc kubenswrapper[4744]: I1008 10:29:47.421276 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzt92" event={"ID":"e3972149-5648-4a77-ae04-1153a8a32565","Type":"ContainerStarted","Data":"c6410e572bbdea15e7581df9d65833f684e97650ad2ce77e334bdbd4d66c42a4"} Oct 08 10:29:58 crc kubenswrapper[4744]: I1008 10:29:58.530963 4744 generic.go:334] "Generic (PLEG): container finished" podID="e3972149-5648-4a77-ae04-1153a8a32565" containerID="c6410e572bbdea15e7581df9d65833f684e97650ad2ce77e334bdbd4d66c42a4" exitCode=0 Oct 08 10:29:58 crc kubenswrapper[4744]: I1008 10:29:58.531015 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzt92" event={"ID":"e3972149-5648-4a77-ae04-1153a8a32565","Type":"ContainerDied","Data":"c6410e572bbdea15e7581df9d65833f684e97650ad2ce77e334bdbd4d66c42a4"} Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.465846 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527"] Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.469019 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.482957 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.485710 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527"] Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.488066 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.634115 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-config-volume\") pod \"collect-profiles-29331990-wn527\" (UID: \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.634292 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4mpz\" (UniqueName: \"kubernetes.io/projected/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-kube-api-access-b4mpz\") pod \"collect-profiles-29331990-wn527\" (UID: \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.634382 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-secret-volume\") pod \"collect-profiles-29331990-wn527\" (UID: \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.757456 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-config-volume\") pod \"collect-profiles-29331990-wn527\" (UID: \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.758224 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4mpz\" (UniqueName: \"kubernetes.io/projected/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-kube-api-access-b4mpz\") pod \"collect-profiles-29331990-wn527\" (UID: \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.758431 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-secret-volume\") pod \"collect-profiles-29331990-wn527\" (UID: \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.761717 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-config-volume\") pod \"collect-profiles-29331990-wn527\" (UID: \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.789584 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4mpz\" (UniqueName: \"kubernetes.io/projected/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-kube-api-access-b4mpz\") pod \"collect-profiles-29331990-wn527\" (UID: \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.791087 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-secret-volume\") pod \"collect-profiles-29331990-wn527\" (UID: \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:00 crc kubenswrapper[4744]: I1008 10:30:00.800706 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:01 crc kubenswrapper[4744]: I1008 10:30:01.268043 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527"] Oct 08 10:30:01 crc kubenswrapper[4744]: I1008 10:30:01.458455 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:30:01 crc kubenswrapper[4744]: E1008 10:30:01.458684 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:30:01 crc kubenswrapper[4744]: I1008 10:30:01.565519 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" event={"ID":"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8","Type":"ContainerStarted","Data":"efef52dbb308ce34fb739202cda709b90ad472caa60df32dfcc3d25ae71c0e5c"} Oct 08 10:30:02 crc kubenswrapper[4744]: I1008 10:30:02.578528 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" event={"ID":"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8","Type":"ContainerStarted","Data":"b4d767cd5c46e4320ce022998367291326fe9212ef79b14714d9b7c7e318f17c"} Oct 08 10:30:02 crc kubenswrapper[4744]: I1008 10:30:02.581283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzt92" event={"ID":"e3972149-5648-4a77-ae04-1153a8a32565","Type":"ContainerStarted","Data":"7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300"} Oct 08 10:30:02 crc kubenswrapper[4744]: I1008 10:30:02.599765 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" podStartSLOduration=2.599733379 podStartE2EDuration="2.599733379s" podCreationTimestamp="2025-10-08 10:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:30:02.594642727 +0000 UTC m=+4697.842287966" watchObservedRunningTime="2025-10-08 10:30:02.599733379 +0000 UTC m=+4697.847378618" Oct 08 10:30:03 crc kubenswrapper[4744]: I1008 10:30:03.590671 4744 generic.go:334] "Generic (PLEG): container finished" podID="3fad4c2d-dfd5-43ed-8597-a3d8831a57d8" containerID="b4d767cd5c46e4320ce022998367291326fe9212ef79b14714d9b7c7e318f17c" exitCode=0 Oct 08 10:30:03 crc kubenswrapper[4744]: I1008 10:30:03.590711 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" event={"ID":"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8","Type":"ContainerDied","Data":"b4d767cd5c46e4320ce022998367291326fe9212ef79b14714d9b7c7e318f17c"} Oct 08 10:30:03 crc kubenswrapper[4744]: I1008 10:30:03.607811 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jzt92" podStartSLOduration=4.305281203 podStartE2EDuration="20.607781052s" podCreationTimestamp="2025-10-08 10:29:43 +0000 UTC" firstStartedPulling="2025-10-08 10:29:45.406405822 +0000 UTC m=+4680.654051051" lastFinishedPulling="2025-10-08 10:30:01.708905661 +0000 UTC m=+4696.956550900" observedRunningTime="2025-10-08 10:30:02.625291668 +0000 UTC m=+4697.872936917" watchObservedRunningTime="2025-10-08 10:30:03.607781052 +0000 UTC m=+4698.855426291" Oct 08 10:30:04 crc kubenswrapper[4744]: I1008 10:30:04.291826 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:30:04 crc kubenswrapper[4744]: I1008 10:30:04.291874 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:30:04 crc kubenswrapper[4744]: I1008 10:30:04.964845 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.071756 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-secret-volume\") pod \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\" (UID: \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\") " Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.071942 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4mpz\" (UniqueName: \"kubernetes.io/projected/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-kube-api-access-b4mpz\") pod \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\" (UID: \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\") " Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.071965 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-config-volume\") pod \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\" (UID: \"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8\") " Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.072990 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-config-volume" (OuterVolumeSpecName: "config-volume") pod "3fad4c2d-dfd5-43ed-8597-a3d8831a57d8" (UID: "3fad4c2d-dfd5-43ed-8597-a3d8831a57d8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.088168 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3fad4c2d-dfd5-43ed-8597-a3d8831a57d8" (UID: "3fad4c2d-dfd5-43ed-8597-a3d8831a57d8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.096773 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-kube-api-access-b4mpz" (OuterVolumeSpecName: "kube-api-access-b4mpz") pod "3fad4c2d-dfd5-43ed-8597-a3d8831a57d8" (UID: "3fad4c2d-dfd5-43ed-8597-a3d8831a57d8"). InnerVolumeSpecName "kube-api-access-b4mpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.174786 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4mpz\" (UniqueName: \"kubernetes.io/projected/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-kube-api-access-b4mpz\") on node \"crc\" DevicePath \"\"" Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.174843 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.174854 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3fad4c2d-dfd5-43ed-8597-a3d8831a57d8-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.346860 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jzt92" podUID="e3972149-5648-4a77-ae04-1153a8a32565" containerName="registry-server" probeResult="failure" output=< Oct 08 10:30:05 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 08 10:30:05 crc kubenswrapper[4744]: > Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.609060 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" event={"ID":"3fad4c2d-dfd5-43ed-8597-a3d8831a57d8","Type":"ContainerDied","Data":"efef52dbb308ce34fb739202cda709b90ad472caa60df32dfcc3d25ae71c0e5c"} Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.609313 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efef52dbb308ce34fb739202cda709b90ad472caa60df32dfcc3d25ae71c0e5c" Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.609094 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331990-wn527" Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.706046 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z"] Oct 08 10:30:05 crc kubenswrapper[4744]: I1008 10:30:05.711999 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331945-8279z"] Oct 08 10:30:07 crc kubenswrapper[4744]: I1008 10:30:07.465553 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d79b1044-6136-4858-93f3-6386a59653f0" path="/var/lib/kubelet/pods/d79b1044-6136-4858-93f3-6386a59653f0/volumes" Oct 08 10:30:12 crc kubenswrapper[4744]: I1008 10:30:12.453407 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:30:12 crc kubenswrapper[4744]: E1008 10:30:12.454196 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:30:14 crc kubenswrapper[4744]: I1008 10:30:14.343656 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:30:14 crc kubenswrapper[4744]: I1008 10:30:14.394324 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:30:15 crc kubenswrapper[4744]: I1008 10:30:15.152333 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jzt92"] Oct 08 10:30:15 crc kubenswrapper[4744]: I1008 10:30:15.695867 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jzt92" podUID="e3972149-5648-4a77-ae04-1153a8a32565" containerName="registry-server" containerID="cri-o://7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300" gracePeriod=2 Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.137870 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.204429 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3972149-5648-4a77-ae04-1153a8a32565-catalog-content\") pod \"e3972149-5648-4a77-ae04-1153a8a32565\" (UID: \"e3972149-5648-4a77-ae04-1153a8a32565\") " Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.204686 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvxct\" (UniqueName: \"kubernetes.io/projected/e3972149-5648-4a77-ae04-1153a8a32565-kube-api-access-nvxct\") pod \"e3972149-5648-4a77-ae04-1153a8a32565\" (UID: \"e3972149-5648-4a77-ae04-1153a8a32565\") " Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.204723 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3972149-5648-4a77-ae04-1153a8a32565-utilities\") pod \"e3972149-5648-4a77-ae04-1153a8a32565\" (UID: \"e3972149-5648-4a77-ae04-1153a8a32565\") " Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.205456 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3972149-5648-4a77-ae04-1153a8a32565-utilities" (OuterVolumeSpecName: "utilities") pod "e3972149-5648-4a77-ae04-1153a8a32565" (UID: "e3972149-5648-4a77-ae04-1153a8a32565"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.209280 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3972149-5648-4a77-ae04-1153a8a32565-kube-api-access-nvxct" (OuterVolumeSpecName: "kube-api-access-nvxct") pod "e3972149-5648-4a77-ae04-1153a8a32565" (UID: "e3972149-5648-4a77-ae04-1153a8a32565"). InnerVolumeSpecName "kube-api-access-nvxct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.289894 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3972149-5648-4a77-ae04-1153a8a32565-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3972149-5648-4a77-ae04-1153a8a32565" (UID: "e3972149-5648-4a77-ae04-1153a8a32565"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.305925 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvxct\" (UniqueName: \"kubernetes.io/projected/e3972149-5648-4a77-ae04-1153a8a32565-kube-api-access-nvxct\") on node \"crc\" DevicePath \"\"" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.305954 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3972149-5648-4a77-ae04-1153a8a32565-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.305963 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3972149-5648-4a77-ae04-1153a8a32565-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.706195 4744 generic.go:334] "Generic (PLEG): container finished" podID="e3972149-5648-4a77-ae04-1153a8a32565" containerID="7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300" exitCode=0 Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.706241 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzt92" event={"ID":"e3972149-5648-4a77-ae04-1153a8a32565","Type":"ContainerDied","Data":"7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300"} Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.706278 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzt92" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.706287 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzt92" event={"ID":"e3972149-5648-4a77-ae04-1153a8a32565","Type":"ContainerDied","Data":"5599f19ee345e7d96d14dc869c77ec209307db9fd975143d108bc7d9fad1f53d"} Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.706304 4744 scope.go:117] "RemoveContainer" containerID="7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.727266 4744 scope.go:117] "RemoveContainer" containerID="c6410e572bbdea15e7581df9d65833f684e97650ad2ce77e334bdbd4d66c42a4" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.745494 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jzt92"] Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.753390 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jzt92"] Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.759041 4744 scope.go:117] "RemoveContainer" containerID="cf8dae41f4dc6c9dc737b675ccdc286868b47d710c677260d0e70a090d00f44e" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.800354 4744 scope.go:117] "RemoveContainer" containerID="7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300" Oct 08 10:30:16 crc kubenswrapper[4744]: E1008 10:30:16.801147 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300\": container with ID starting with 7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300 not found: ID does not exist" containerID="7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.801190 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300"} err="failed to get container status \"7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300\": rpc error: code = NotFound desc = could not find container \"7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300\": container with ID starting with 7ae628b71558e2e7425ccf193626ae72992f508e8ca06d61b4a265409378c300 not found: ID does not exist" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.801226 4744 scope.go:117] "RemoveContainer" containerID="c6410e572bbdea15e7581df9d65833f684e97650ad2ce77e334bdbd4d66c42a4" Oct 08 10:30:16 crc kubenswrapper[4744]: E1008 10:30:16.801532 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6410e572bbdea15e7581df9d65833f684e97650ad2ce77e334bdbd4d66c42a4\": container with ID starting with c6410e572bbdea15e7581df9d65833f684e97650ad2ce77e334bdbd4d66c42a4 not found: ID does not exist" containerID="c6410e572bbdea15e7581df9d65833f684e97650ad2ce77e334bdbd4d66c42a4" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.801562 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6410e572bbdea15e7581df9d65833f684e97650ad2ce77e334bdbd4d66c42a4"} err="failed to get container status \"c6410e572bbdea15e7581df9d65833f684e97650ad2ce77e334bdbd4d66c42a4\": rpc error: code = NotFound desc = could not find container \"c6410e572bbdea15e7581df9d65833f684e97650ad2ce77e334bdbd4d66c42a4\": container with ID starting with c6410e572bbdea15e7581df9d65833f684e97650ad2ce77e334bdbd4d66c42a4 not found: ID does not exist" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.801602 4744 scope.go:117] "RemoveContainer" containerID="cf8dae41f4dc6c9dc737b675ccdc286868b47d710c677260d0e70a090d00f44e" Oct 08 10:30:16 crc kubenswrapper[4744]: E1008 10:30:16.801866 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf8dae41f4dc6c9dc737b675ccdc286868b47d710c677260d0e70a090d00f44e\": container with ID starting with cf8dae41f4dc6c9dc737b675ccdc286868b47d710c677260d0e70a090d00f44e not found: ID does not exist" containerID="cf8dae41f4dc6c9dc737b675ccdc286868b47d710c677260d0e70a090d00f44e" Oct 08 10:30:16 crc kubenswrapper[4744]: I1008 10:30:16.801904 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf8dae41f4dc6c9dc737b675ccdc286868b47d710c677260d0e70a090d00f44e"} err="failed to get container status \"cf8dae41f4dc6c9dc737b675ccdc286868b47d710c677260d0e70a090d00f44e\": rpc error: code = NotFound desc = could not find container \"cf8dae41f4dc6c9dc737b675ccdc286868b47d710c677260d0e70a090d00f44e\": container with ID starting with cf8dae41f4dc6c9dc737b675ccdc286868b47d710c677260d0e70a090d00f44e not found: ID does not exist" Oct 08 10:30:17 crc kubenswrapper[4744]: I1008 10:30:17.465912 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3972149-5648-4a77-ae04-1153a8a32565" path="/var/lib/kubelet/pods/e3972149-5648-4a77-ae04-1153a8a32565/volumes" Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.817570 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-q6d7g/must-gather-f52xr"] Oct 08 10:30:19 crc kubenswrapper[4744]: E1008 10:30:19.818482 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3972149-5648-4a77-ae04-1153a8a32565" containerName="extract-utilities" Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.818497 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3972149-5648-4a77-ae04-1153a8a32565" containerName="extract-utilities" Oct 08 10:30:19 crc kubenswrapper[4744]: E1008 10:30:19.818526 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3972149-5648-4a77-ae04-1153a8a32565" containerName="registry-server" Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.818532 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3972149-5648-4a77-ae04-1153a8a32565" containerName="registry-server" Oct 08 10:30:19 crc kubenswrapper[4744]: E1008 10:30:19.818552 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3972149-5648-4a77-ae04-1153a8a32565" containerName="extract-content" Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.818558 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3972149-5648-4a77-ae04-1153a8a32565" containerName="extract-content" Oct 08 10:30:19 crc kubenswrapper[4744]: E1008 10:30:19.818574 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fad4c2d-dfd5-43ed-8597-a3d8831a57d8" containerName="collect-profiles" Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.818580 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fad4c2d-dfd5-43ed-8597-a3d8831a57d8" containerName="collect-profiles" Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.818741 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fad4c2d-dfd5-43ed-8597-a3d8831a57d8" containerName="collect-profiles" Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.818751 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3972149-5648-4a77-ae04-1153a8a32565" containerName="registry-server" Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.819730 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/must-gather-f52xr" Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.828171 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-q6d7g/must-gather-f52xr"] Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.835831 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-q6d7g"/"openshift-service-ca.crt" Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.836005 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-q6d7g"/"kube-root-ca.crt" Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.981290 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djm7m\" (UniqueName: \"kubernetes.io/projected/de13d6e4-3524-4157-837a-eec911bc5316-kube-api-access-djm7m\") pod \"must-gather-f52xr\" (UID: \"de13d6e4-3524-4157-837a-eec911bc5316\") " pod="openshift-must-gather-q6d7g/must-gather-f52xr" Oct 08 10:30:19 crc kubenswrapper[4744]: I1008 10:30:19.981416 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/de13d6e4-3524-4157-837a-eec911bc5316-must-gather-output\") pod \"must-gather-f52xr\" (UID: \"de13d6e4-3524-4157-837a-eec911bc5316\") " pod="openshift-must-gather-q6d7g/must-gather-f52xr" Oct 08 10:30:20 crc kubenswrapper[4744]: I1008 10:30:20.082877 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djm7m\" (UniqueName: \"kubernetes.io/projected/de13d6e4-3524-4157-837a-eec911bc5316-kube-api-access-djm7m\") pod \"must-gather-f52xr\" (UID: \"de13d6e4-3524-4157-837a-eec911bc5316\") " pod="openshift-must-gather-q6d7g/must-gather-f52xr" Oct 08 10:30:20 crc kubenswrapper[4744]: I1008 10:30:20.082965 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/de13d6e4-3524-4157-837a-eec911bc5316-must-gather-output\") pod \"must-gather-f52xr\" (UID: \"de13d6e4-3524-4157-837a-eec911bc5316\") " pod="openshift-must-gather-q6d7g/must-gather-f52xr" Oct 08 10:30:20 crc kubenswrapper[4744]: I1008 10:30:20.083547 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/de13d6e4-3524-4157-837a-eec911bc5316-must-gather-output\") pod \"must-gather-f52xr\" (UID: \"de13d6e4-3524-4157-837a-eec911bc5316\") " pod="openshift-must-gather-q6d7g/must-gather-f52xr" Oct 08 10:30:20 crc kubenswrapper[4744]: I1008 10:30:20.101972 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djm7m\" (UniqueName: \"kubernetes.io/projected/de13d6e4-3524-4157-837a-eec911bc5316-kube-api-access-djm7m\") pod \"must-gather-f52xr\" (UID: \"de13d6e4-3524-4157-837a-eec911bc5316\") " pod="openshift-must-gather-q6d7g/must-gather-f52xr" Oct 08 10:30:20 crc kubenswrapper[4744]: I1008 10:30:20.154756 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/must-gather-f52xr" Oct 08 10:30:20 crc kubenswrapper[4744]: I1008 10:30:20.715830 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-q6d7g/must-gather-f52xr"] Oct 08 10:30:20 crc kubenswrapper[4744]: W1008 10:30:20.737543 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde13d6e4_3524_4157_837a_eec911bc5316.slice/crio-3e267251b7cc2ecab3e34e65b8a4a24bb66f86c8c60f45d92b028fe1ade9b4b4 WatchSource:0}: Error finding container 3e267251b7cc2ecab3e34e65b8a4a24bb66f86c8c60f45d92b028fe1ade9b4b4: Status 404 returned error can't find the container with id 3e267251b7cc2ecab3e34e65b8a4a24bb66f86c8c60f45d92b028fe1ade9b4b4 Oct 08 10:30:20 crc kubenswrapper[4744]: I1008 10:30:20.764382 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q6d7g/must-gather-f52xr" event={"ID":"de13d6e4-3524-4157-837a-eec911bc5316","Type":"ContainerStarted","Data":"3e267251b7cc2ecab3e34e65b8a4a24bb66f86c8c60f45d92b028fe1ade9b4b4"} Oct 08 10:30:22 crc kubenswrapper[4744]: I1008 10:30:22.781955 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q6d7g/must-gather-f52xr" event={"ID":"de13d6e4-3524-4157-837a-eec911bc5316","Type":"ContainerStarted","Data":"54cb51c42a803df0b907d6d7cec344c30b0ca2884516490478d601c9fbbb6b0e"} Oct 08 10:30:22 crc kubenswrapper[4744]: I1008 10:30:22.782464 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q6d7g/must-gather-f52xr" event={"ID":"de13d6e4-3524-4157-837a-eec911bc5316","Type":"ContainerStarted","Data":"27b64049af580689f69aab067417d368043c7d29952c292ba8476841a92abace"} Oct 08 10:30:22 crc kubenswrapper[4744]: I1008 10:30:22.796127 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-q6d7g/must-gather-f52xr" podStartSLOduration=3.79610876 podStartE2EDuration="3.79610876s" podCreationTimestamp="2025-10-08 10:30:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:30:22.793629991 +0000 UTC m=+4718.041275250" watchObservedRunningTime="2025-10-08 10:30:22.79610876 +0000 UTC m=+4718.043753999" Oct 08 10:30:25 crc kubenswrapper[4744]: I1008 10:30:25.460015 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:30:25 crc kubenswrapper[4744]: E1008 10:30:25.460704 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:30:26 crc kubenswrapper[4744]: I1008 10:30:26.496030 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-q6d7g/crc-debug-7jhxr"] Oct 08 10:30:26 crc kubenswrapper[4744]: I1008 10:30:26.497634 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" Oct 08 10:30:26 crc kubenswrapper[4744]: I1008 10:30:26.503094 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-q6d7g"/"default-dockercfg-88nh5" Oct 08 10:30:26 crc kubenswrapper[4744]: I1008 10:30:26.614259 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwrv2\" (UniqueName: \"kubernetes.io/projected/6fc6b4e7-f371-45e6-9e2b-307e009a2f86-kube-api-access-mwrv2\") pod \"crc-debug-7jhxr\" (UID: \"6fc6b4e7-f371-45e6-9e2b-307e009a2f86\") " pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" Oct 08 10:30:26 crc kubenswrapper[4744]: I1008 10:30:26.614686 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6fc6b4e7-f371-45e6-9e2b-307e009a2f86-host\") pod \"crc-debug-7jhxr\" (UID: \"6fc6b4e7-f371-45e6-9e2b-307e009a2f86\") " pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" Oct 08 10:30:26 crc kubenswrapper[4744]: I1008 10:30:26.716396 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6fc6b4e7-f371-45e6-9e2b-307e009a2f86-host\") pod \"crc-debug-7jhxr\" (UID: \"6fc6b4e7-f371-45e6-9e2b-307e009a2f86\") " pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" Oct 08 10:30:26 crc kubenswrapper[4744]: I1008 10:30:26.716489 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwrv2\" (UniqueName: \"kubernetes.io/projected/6fc6b4e7-f371-45e6-9e2b-307e009a2f86-kube-api-access-mwrv2\") pod \"crc-debug-7jhxr\" (UID: \"6fc6b4e7-f371-45e6-9e2b-307e009a2f86\") " pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" Oct 08 10:30:26 crc kubenswrapper[4744]: I1008 10:30:26.716571 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6fc6b4e7-f371-45e6-9e2b-307e009a2f86-host\") pod \"crc-debug-7jhxr\" (UID: \"6fc6b4e7-f371-45e6-9e2b-307e009a2f86\") " pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" Oct 08 10:30:26 crc kubenswrapper[4744]: I1008 10:30:26.905479 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwrv2\" (UniqueName: \"kubernetes.io/projected/6fc6b4e7-f371-45e6-9e2b-307e009a2f86-kube-api-access-mwrv2\") pod \"crc-debug-7jhxr\" (UID: \"6fc6b4e7-f371-45e6-9e2b-307e009a2f86\") " pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" Oct 08 10:30:27 crc kubenswrapper[4744]: I1008 10:30:27.116952 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" Oct 08 10:30:27 crc kubenswrapper[4744]: I1008 10:30:27.828984 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" event={"ID":"6fc6b4e7-f371-45e6-9e2b-307e009a2f86","Type":"ContainerStarted","Data":"2aecdb27ecc55d2d48e80ccf933111d11798332b0a2358d121aa8581d3ac1314"} Oct 08 10:30:27 crc kubenswrapper[4744]: I1008 10:30:27.829523 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" event={"ID":"6fc6b4e7-f371-45e6-9e2b-307e009a2f86","Type":"ContainerStarted","Data":"40fc7efa689dae81cfbd8bab027178b498876ff90ddbc8daf10e8628fb5b6dee"} Oct 08 10:30:27 crc kubenswrapper[4744]: I1008 10:30:27.843390 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" podStartSLOduration=1.8433489010000002 podStartE2EDuration="1.843348901s" podCreationTimestamp="2025-10-08 10:30:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 10:30:27.839490404 +0000 UTC m=+4723.087135643" watchObservedRunningTime="2025-10-08 10:30:27.843348901 +0000 UTC m=+4723.090994140" Oct 08 10:30:38 crc kubenswrapper[4744]: I1008 10:30:38.453408 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:30:38 crc kubenswrapper[4744]: E1008 10:30:38.454941 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:30:51 crc kubenswrapper[4744]: I1008 10:30:51.453675 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:30:52 crc kubenswrapper[4744]: I1008 10:30:52.028730 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"54379f58e3765719921fb408b4441ea4d56d9cecce1a289805e1d0aad427c6de"} Oct 08 10:30:59 crc kubenswrapper[4744]: I1008 10:30:59.525909 4744 scope.go:117] "RemoveContainer" containerID="b46fb2beb2af63fcd68ed6b1414c4c5f32f8f019e1c4533ccb16e3db5515dc32" Oct 08 10:30:59 crc kubenswrapper[4744]: I1008 10:30:59.553841 4744 scope.go:117] "RemoveContainer" containerID="2db52d6c8e769a1197b7e6367f657bdbef3be33119fc78af04557521a5c5a858" Oct 08 10:32:19 crc kubenswrapper[4744]: I1008 10:32:19.891855 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6d98c4654d-2xrfx_ea9a2551-0abd-4bce-9785-4c8c7e11b02d/barbican-api/0.log" Oct 08 10:32:20 crc kubenswrapper[4744]: I1008 10:32:20.024839 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6d98c4654d-2xrfx_ea9a2551-0abd-4bce-9785-4c8c7e11b02d/barbican-api-log/0.log" Oct 08 10:32:20 crc kubenswrapper[4744]: I1008 10:32:20.350463 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-8578895ff6-xg92r_4ca73575-280f-43af-8c34-123a1cf83e92/barbican-keystone-listener/0.log" Oct 08 10:32:20 crc kubenswrapper[4744]: I1008 10:32:20.422971 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-8578895ff6-xg92r_4ca73575-280f-43af-8c34-123a1cf83e92/barbican-keystone-listener-log/0.log" Oct 08 10:32:20 crc kubenswrapper[4744]: I1008 10:32:20.616717 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5ccc6f877f-j2zc9_203b8276-cafa-4b89-a68d-0f9437ee7161/barbican-worker/0.log" Oct 08 10:32:20 crc kubenswrapper[4744]: I1008 10:32:20.759398 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5ccc6f877f-j2zc9_203b8276-cafa-4b89-a68d-0f9437ee7161/barbican-worker-log/0.log" Oct 08 10:32:21 crc kubenswrapper[4744]: I1008 10:32:21.170986 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-dntq4_f0464e55-2b52-4c5a-bc87-9c2c87cb6101/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:21 crc kubenswrapper[4744]: I1008 10:32:21.297017 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2568ab3e-d5f6-4524-b1dc-20a6f389cd31/ceilometer-central-agent/0.log" Oct 08 10:32:21 crc kubenswrapper[4744]: I1008 10:32:21.381273 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2568ab3e-d5f6-4524-b1dc-20a6f389cd31/ceilometer-notification-agent/0.log" Oct 08 10:32:21 crc kubenswrapper[4744]: I1008 10:32:21.479037 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2568ab3e-d5f6-4524-b1dc-20a6f389cd31/proxy-httpd/0.log" Oct 08 10:32:21 crc kubenswrapper[4744]: I1008 10:32:21.541306 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_2568ab3e-d5f6-4524-b1dc-20a6f389cd31/sg-core/0.log" Oct 08 10:32:21 crc kubenswrapper[4744]: I1008 10:32:21.885438 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-sw54k_664b4c04-3a69-444f-aa77-621229ee5b96/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:22 crc kubenswrapper[4744]: I1008 10:32:22.042403 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-hklgl_dd384bed-7c22-4e1a-8342-94589991c934/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:22 crc kubenswrapper[4744]: I1008 10:32:22.248979 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9efda746-3b68-4707-bbe8-3f4e74db5a60/cinder-api/0.log" Oct 08 10:32:22 crc kubenswrapper[4744]: I1008 10:32:22.409343 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9efda746-3b68-4707-bbe8-3f4e74db5a60/cinder-api-log/0.log" Oct 08 10:32:22 crc kubenswrapper[4744]: I1008 10:32:22.775286 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_d4fd8637-8969-4c7f-9ecf-9276b5550484/probe/0.log" Oct 08 10:32:22 crc kubenswrapper[4744]: I1008 10:32:22.799362 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_d4fd8637-8969-4c7f-9ecf-9276b5550484/cinder-backup/0.log" Oct 08 10:32:23 crc kubenswrapper[4744]: I1008 10:32:23.270953 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_49177d36-177b-4b09-a93e-3d602ff30a27/probe/0.log" Oct 08 10:32:23 crc kubenswrapper[4744]: I1008 10:32:23.288299 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_49177d36-177b-4b09-a93e-3d602ff30a27/cinder-scheduler/0.log" Oct 08 10:32:23 crc kubenswrapper[4744]: I1008 10:32:23.530722 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_9cc1c714-af01-4205-8670-9055cdb3c623/probe/0.log" Oct 08 10:32:23 crc kubenswrapper[4744]: I1008 10:32:23.554100 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_9cc1c714-af01-4205-8670-9055cdb3c623/cinder-volume/0.log" Oct 08 10:32:23 crc kubenswrapper[4744]: I1008 10:32:23.673943 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-k48qv_d0fc8259-e8cd-4d69-8a95-f74d908d569e/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:23 crc kubenswrapper[4744]: I1008 10:32:23.838138 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-wcf57_bf19b11f-c43e-4f3b-a34c-53d3837705fb/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:23 crc kubenswrapper[4744]: I1008 10:32:23.914071 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7887c4559f-jqcwg_084ffd95-c3f3-4340-bc26-d3c2c427c8ad/init/0.log" Oct 08 10:32:24 crc kubenswrapper[4744]: I1008 10:32:24.152828 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7887c4559f-jqcwg_084ffd95-c3f3-4340-bc26-d3c2c427c8ad/init/0.log" Oct 08 10:32:24 crc kubenswrapper[4744]: I1008 10:32:24.274584 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_26c9e730-4ef5-41f3-b861-fbe778351306/glance-httpd/0.log" Oct 08 10:32:24 crc kubenswrapper[4744]: I1008 10:32:24.396463 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7887c4559f-jqcwg_084ffd95-c3f3-4340-bc26-d3c2c427c8ad/dnsmasq-dns/0.log" Oct 08 10:32:24 crc kubenswrapper[4744]: I1008 10:32:24.506909 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_26c9e730-4ef5-41f3-b861-fbe778351306/glance-log/0.log" Oct 08 10:32:24 crc kubenswrapper[4744]: I1008 10:32:24.550047 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2bdd463c-e43b-4740-ba42-3981cb6b1380/glance-httpd/0.log" Oct 08 10:32:24 crc kubenswrapper[4744]: I1008 10:32:24.618191 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_2bdd463c-e43b-4740-ba42-3981cb6b1380/glance-log/0.log" Oct 08 10:32:24 crc kubenswrapper[4744]: I1008 10:32:24.873832 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5549798486-rbbgm_f337fa5f-1088-407d-a832-856ff504cd72/horizon/0.log" Oct 08 10:32:25 crc kubenswrapper[4744]: I1008 10:32:25.042636 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5549798486-rbbgm_f337fa5f-1088-407d-a832-856ff504cd72/horizon-log/0.log" Oct 08 10:32:25 crc kubenswrapper[4744]: I1008 10:32:25.153674 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bctfd_3876b90a-308b-4320-a8ab-f81a48295c1f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:25 crc kubenswrapper[4744]: I1008 10:32:25.403095 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-cd6ks_57e3beb8-aeaa-435a-aa5a-78f854db1d6d/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:25 crc kubenswrapper[4744]: I1008 10:32:25.574050 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c4d7b4588-6wcr4_2af8995b-abf6-4615-bfda-f5900a1bfb2c/keystone-api/0.log" Oct 08 10:32:25 crc kubenswrapper[4744]: I1008 10:32:25.619932 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29331961-k29hs_6d3e0b57-177d-4351-8fce-3a758311bb35/keystone-cron/0.log" Oct 08 10:32:25 crc kubenswrapper[4744]: I1008 10:32:25.736652 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a8b493e3-c187-4150-95db-3a4f110d15ea/kube-state-metrics/0.log" Oct 08 10:32:25 crc kubenswrapper[4744]: I1008 10:32:25.964247 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-swsrc_e80d5289-7657-4835-8b43-936537129935/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:26 crc kubenswrapper[4744]: I1008 10:32:26.143105 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_1003e3fc-33a3-45d1-a1b7-9f383af6e336/manila-api-log/0.log" Oct 08 10:32:26 crc kubenswrapper[4744]: I1008 10:32:26.164833 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_1003e3fc-33a3-45d1-a1b7-9f383af6e336/manila-api/0.log" Oct 08 10:32:26 crc kubenswrapper[4744]: I1008 10:32:26.287095 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e/manila-scheduler/0.log" Oct 08 10:32:26 crc kubenswrapper[4744]: I1008 10:32:26.349330 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_3d0e7dd3-3ec3-499b-abd9-85d2333b5e6e/probe/0.log" Oct 08 10:32:26 crc kubenswrapper[4744]: I1008 10:32:26.537793 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_0e308a1c-d62d-4782-adba-7517aab54f94/probe/0.log" Oct 08 10:32:26 crc kubenswrapper[4744]: I1008 10:32:26.621666 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_0e308a1c-d62d-4782-adba-7517aab54f94/manila-share/0.log" Oct 08 10:32:26 crc kubenswrapper[4744]: I1008 10:32:26.979911 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6b4cb6945c-kjx6m_d7cda27f-1e6a-4637-b647-54f6a8235ce0/neutron-httpd/0.log" Oct 08 10:32:27 crc kubenswrapper[4744]: I1008 10:32:27.090054 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6b4cb6945c-kjx6m_d7cda27f-1e6a-4637-b647-54f6a8235ce0/neutron-api/0.log" Oct 08 10:32:27 crc kubenswrapper[4744]: I1008 10:32:27.763967 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-9c7wx_3f2cc21c-f2e1-4775-8563-c794073928f4/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:28 crc kubenswrapper[4744]: I1008 10:32:28.449144 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_63d92921-2a92-4233-8f69-02aa2c2b5984/nova-api-log/0.log" Oct 08 10:32:28 crc kubenswrapper[4744]: I1008 10:32:28.774576 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_3e036c58-fa76-41bb-a16f-77566f7e5533/nova-cell0-conductor-conductor/0.log" Oct 08 10:32:28 crc kubenswrapper[4744]: I1008 10:32:28.862993 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_63d92921-2a92-4233-8f69-02aa2c2b5984/nova-api-api/0.log" Oct 08 10:32:29 crc kubenswrapper[4744]: I1008 10:32:29.562153 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0b03c126-e043-4a2e-ae42-e9a359622fcd/nova-cell1-conductor-conductor/0.log" Oct 08 10:32:29 crc kubenswrapper[4744]: I1008 10:32:29.593512 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_a197fde0-ffb1-4ead-bdd8-acd2e60bc97f/nova-cell1-novncproxy-novncproxy/0.log" Oct 08 10:32:30 crc kubenswrapper[4744]: I1008 10:32:30.086637 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0592f7bc-38a5-43ad-ae49-acdb4f9023f4/nova-metadata-log/0.log" Oct 08 10:32:30 crc kubenswrapper[4744]: I1008 10:32:30.122580 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-s5v8z_43d8aacb-0649-483b-9e07-bde175fd9d52/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:30 crc kubenswrapper[4744]: I1008 10:32:30.710499 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ee6ac453-8ad5-4c02-adf0-ff88a946dd15/nova-scheduler-scheduler/0.log" Oct 08 10:32:30 crc kubenswrapper[4744]: I1008 10:32:30.928036 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ef3dc460-b269-4e4c-b1db-bfb550b58a7b/mysql-bootstrap/0.log" Oct 08 10:32:31 crc kubenswrapper[4744]: I1008 10:32:31.179273 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ef3dc460-b269-4e4c-b1db-bfb550b58a7b/mysql-bootstrap/0.log" Oct 08 10:32:31 crc kubenswrapper[4744]: I1008 10:32:31.206917 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ef3dc460-b269-4e4c-b1db-bfb550b58a7b/galera/0.log" Oct 08 10:32:31 crc kubenswrapper[4744]: I1008 10:32:31.519394 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d1412b08-0e26-455c-afb5-3f51b2cb3012/mysql-bootstrap/0.log" Oct 08 10:32:31 crc kubenswrapper[4744]: I1008 10:32:31.643513 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d1412b08-0e26-455c-afb5-3f51b2cb3012/mysql-bootstrap/0.log" Oct 08 10:32:31 crc kubenswrapper[4744]: I1008 10:32:31.740528 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d1412b08-0e26-455c-afb5-3f51b2cb3012/galera/0.log" Oct 08 10:32:31 crc kubenswrapper[4744]: I1008 10:32:31.944788 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6fd3ee7f-6102-426d-b482-a624a7871f42/openstackclient/0.log" Oct 08 10:32:32 crc kubenswrapper[4744]: I1008 10:32:32.084208 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0592f7bc-38a5-43ad-ae49-acdb4f9023f4/nova-metadata-metadata/0.log" Oct 08 10:32:32 crc kubenswrapper[4744]: I1008 10:32:32.208347 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kshz8_82e01314-590d-4fe4-9bbb-e5d180ff7093/openstack-network-exporter/0.log" Oct 08 10:32:32 crc kubenswrapper[4744]: I1008 10:32:32.500490 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-mhqr7_ad8606dc-c6d0-4ae7-905e-929d0a82ae8d/ovn-controller/0.log" Oct 08 10:32:32 crc kubenswrapper[4744]: I1008 10:32:32.718787 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ff8dm_fef2319a-5415-4712-a7c6-ad0ab4decff0/ovsdb-server-init/0.log" Oct 08 10:32:33 crc kubenswrapper[4744]: I1008 10:32:33.016290 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ff8dm_fef2319a-5415-4712-a7c6-ad0ab4decff0/ovs-vswitchd/0.log" Oct 08 10:32:33 crc kubenswrapper[4744]: I1008 10:32:33.081592 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ff8dm_fef2319a-5415-4712-a7c6-ad0ab4decff0/ovsdb-server-init/0.log" Oct 08 10:32:33 crc kubenswrapper[4744]: I1008 10:32:33.118846 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-ff8dm_fef2319a-5415-4712-a7c6-ad0ab4decff0/ovsdb-server/0.log" Oct 08 10:32:33 crc kubenswrapper[4744]: I1008 10:32:33.467637 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-qslf4_fa1452ca-c8a1-4185-b5ee-9119851d2e20/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:33 crc kubenswrapper[4744]: I1008 10:32:33.760043 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2c0cba37-bef9-4b03-8c6a-43acd36d04ff/ovn-northd/0.log" Oct 08 10:32:33 crc kubenswrapper[4744]: I1008 10:32:33.777608 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_2c0cba37-bef9-4b03-8c6a-43acd36d04ff/openstack-network-exporter/0.log" Oct 08 10:32:34 crc kubenswrapper[4744]: I1008 10:32:34.031475 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a38f1196-55fe-462c-b012-be0d5fbe33ee/openstack-network-exporter/0.log" Oct 08 10:32:34 crc kubenswrapper[4744]: I1008 10:32:34.100102 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a38f1196-55fe-462c-b012-be0d5fbe33ee/ovsdbserver-nb/0.log" Oct 08 10:32:34 crc kubenswrapper[4744]: I1008 10:32:34.314576 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_815ecda8-d0f0-47c1-b638-11a5ca9fce02/openstack-network-exporter/0.log" Oct 08 10:32:34 crc kubenswrapper[4744]: I1008 10:32:34.373608 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_815ecda8-d0f0-47c1-b638-11a5ca9fce02/ovsdbserver-sb/0.log" Oct 08 10:32:34 crc kubenswrapper[4744]: I1008 10:32:34.734218 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7f48465966-bh5rl_79381d24-0ce2-46ee-ae85-6fdbcbbae230/placement-api/0.log" Oct 08 10:32:34 crc kubenswrapper[4744]: I1008 10:32:34.870532 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7f48465966-bh5rl_79381d24-0ce2-46ee-ae85-6fdbcbbae230/placement-log/0.log" Oct 08 10:32:35 crc kubenswrapper[4744]: I1008 10:32:35.007617 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8ae09ae5-011d-48f9-ad83-a53bfb91c90c/setup-container/0.log" Oct 08 10:32:35 crc kubenswrapper[4744]: I1008 10:32:35.273006 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8ae09ae5-011d-48f9-ad83-a53bfb91c90c/setup-container/0.log" Oct 08 10:32:35 crc kubenswrapper[4744]: I1008 10:32:35.403698 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8ae09ae5-011d-48f9-ad83-a53bfb91c90c/rabbitmq/0.log" Oct 08 10:32:35 crc kubenswrapper[4744]: I1008 10:32:35.527397 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_19186b52-dfab-4d1f-9ccd-176abc9975ff/setup-container/0.log" Oct 08 10:32:35 crc kubenswrapper[4744]: I1008 10:32:35.894660 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_19186b52-dfab-4d1f-9ccd-176abc9975ff/rabbitmq/0.log" Oct 08 10:32:35 crc kubenswrapper[4744]: I1008 10:32:35.968143 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_19186b52-dfab-4d1f-9ccd-176abc9975ff/setup-container/0.log" Oct 08 10:32:36 crc kubenswrapper[4744]: I1008 10:32:36.110797 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8smg2_446dc944-bd86-4c7a-a9da-bc1193d34d33/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:36 crc kubenswrapper[4744]: I1008 10:32:36.312456 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-l8d96_50611401-ad48-4b64-ad34-48a7f2d71a11/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:36 crc kubenswrapper[4744]: I1008 10:32:36.610144 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-jmbgg_77eb2e7f-5360-4589-bb9f-56d6a28fc3b0/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:37 crc kubenswrapper[4744]: I1008 10:32:37.183441 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-f6q5g_ed43f66e-c9f8-4de0-9604-f5a1b993ed03/ssh-known-hosts-edpm-deployment/0.log" Oct 08 10:32:37 crc kubenswrapper[4744]: I1008 10:32:37.418218 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_ba86d64c-cc8a-48b2-b70e-2ca6cf2145be/tempest-tests-tempest-tests-runner/0.log" Oct 08 10:32:37 crc kubenswrapper[4744]: I1008 10:32:37.572148 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_e36a0230-d350-400e-a974-ce78f036dd64/test-operator-logs-container/0.log" Oct 08 10:32:37 crc kubenswrapper[4744]: I1008 10:32:37.667711 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-6nkzh_4e3fa5bc-5c8b-45de-a173-daf069f2868f/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 10:32:38 crc kubenswrapper[4744]: I1008 10:32:38.115353 4744 generic.go:334] "Generic (PLEG): container finished" podID="6fc6b4e7-f371-45e6-9e2b-307e009a2f86" containerID="2aecdb27ecc55d2d48e80ccf933111d11798332b0a2358d121aa8581d3ac1314" exitCode=0 Oct 08 10:32:38 crc kubenswrapper[4744]: I1008 10:32:38.115721 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" event={"ID":"6fc6b4e7-f371-45e6-9e2b-307e009a2f86","Type":"ContainerDied","Data":"2aecdb27ecc55d2d48e80ccf933111d11798332b0a2358d121aa8581d3ac1314"} Oct 08 10:32:39 crc kubenswrapper[4744]: I1008 10:32:39.227987 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" Oct 08 10:32:39 crc kubenswrapper[4744]: I1008 10:32:39.270415 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-q6d7g/crc-debug-7jhxr"] Oct 08 10:32:39 crc kubenswrapper[4744]: I1008 10:32:39.277362 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-q6d7g/crc-debug-7jhxr"] Oct 08 10:32:39 crc kubenswrapper[4744]: I1008 10:32:39.352221 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwrv2\" (UniqueName: \"kubernetes.io/projected/6fc6b4e7-f371-45e6-9e2b-307e009a2f86-kube-api-access-mwrv2\") pod \"6fc6b4e7-f371-45e6-9e2b-307e009a2f86\" (UID: \"6fc6b4e7-f371-45e6-9e2b-307e009a2f86\") " Oct 08 10:32:39 crc kubenswrapper[4744]: I1008 10:32:39.352313 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6fc6b4e7-f371-45e6-9e2b-307e009a2f86-host\") pod \"6fc6b4e7-f371-45e6-9e2b-307e009a2f86\" (UID: \"6fc6b4e7-f371-45e6-9e2b-307e009a2f86\") " Oct 08 10:32:39 crc kubenswrapper[4744]: I1008 10:32:39.352397 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6fc6b4e7-f371-45e6-9e2b-307e009a2f86-host" (OuterVolumeSpecName: "host") pod "6fc6b4e7-f371-45e6-9e2b-307e009a2f86" (UID: "6fc6b4e7-f371-45e6-9e2b-307e009a2f86"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 10:32:39 crc kubenswrapper[4744]: I1008 10:32:39.352830 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6fc6b4e7-f371-45e6-9e2b-307e009a2f86-host\") on node \"crc\" DevicePath \"\"" Oct 08 10:32:39 crc kubenswrapper[4744]: I1008 10:32:39.387634 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fc6b4e7-f371-45e6-9e2b-307e009a2f86-kube-api-access-mwrv2" (OuterVolumeSpecName: "kube-api-access-mwrv2") pod "6fc6b4e7-f371-45e6-9e2b-307e009a2f86" (UID: "6fc6b4e7-f371-45e6-9e2b-307e009a2f86"). InnerVolumeSpecName "kube-api-access-mwrv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:32:39 crc kubenswrapper[4744]: I1008 10:32:39.454220 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwrv2\" (UniqueName: \"kubernetes.io/projected/6fc6b4e7-f371-45e6-9e2b-307e009a2f86-kube-api-access-mwrv2\") on node \"crc\" DevicePath \"\"" Oct 08 10:32:39 crc kubenswrapper[4744]: I1008 10:32:39.468139 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fc6b4e7-f371-45e6-9e2b-307e009a2f86" path="/var/lib/kubelet/pods/6fc6b4e7-f371-45e6-9e2b-307e009a2f86/volumes" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.133576 4744 scope.go:117] "RemoveContainer" containerID="2aecdb27ecc55d2d48e80ccf933111d11798332b0a2358d121aa8581d3ac1314" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.133875 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/crc-debug-7jhxr" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.554056 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-q6d7g/crc-debug-v896p"] Oct 08 10:32:40 crc kubenswrapper[4744]: E1008 10:32:40.559713 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fc6b4e7-f371-45e6-9e2b-307e009a2f86" containerName="container-00" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.559739 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fc6b4e7-f371-45e6-9e2b-307e009a2f86" containerName="container-00" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.559930 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fc6b4e7-f371-45e6-9e2b-307e009a2f86" containerName="container-00" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.560813 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/crc-debug-v896p" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.567669 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-q6d7g"/"default-dockercfg-88nh5" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.676787 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dfa370e6-e90e-4a55-b8d0-332f370999b1-host\") pod \"crc-debug-v896p\" (UID: \"dfa370e6-e90e-4a55-b8d0-332f370999b1\") " pod="openshift-must-gather-q6d7g/crc-debug-v896p" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.677747 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwrdj\" (UniqueName: \"kubernetes.io/projected/dfa370e6-e90e-4a55-b8d0-332f370999b1-kube-api-access-zwrdj\") pod \"crc-debug-v896p\" (UID: \"dfa370e6-e90e-4a55-b8d0-332f370999b1\") " pod="openshift-must-gather-q6d7g/crc-debug-v896p" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.779095 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwrdj\" (UniqueName: \"kubernetes.io/projected/dfa370e6-e90e-4a55-b8d0-332f370999b1-kube-api-access-zwrdj\") pod \"crc-debug-v896p\" (UID: \"dfa370e6-e90e-4a55-b8d0-332f370999b1\") " pod="openshift-must-gather-q6d7g/crc-debug-v896p" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.779716 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dfa370e6-e90e-4a55-b8d0-332f370999b1-host\") pod \"crc-debug-v896p\" (UID: \"dfa370e6-e90e-4a55-b8d0-332f370999b1\") " pod="openshift-must-gather-q6d7g/crc-debug-v896p" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.779951 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dfa370e6-e90e-4a55-b8d0-332f370999b1-host\") pod \"crc-debug-v896p\" (UID: \"dfa370e6-e90e-4a55-b8d0-332f370999b1\") " pod="openshift-must-gather-q6d7g/crc-debug-v896p" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.802419 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwrdj\" (UniqueName: \"kubernetes.io/projected/dfa370e6-e90e-4a55-b8d0-332f370999b1-kube-api-access-zwrdj\") pod \"crc-debug-v896p\" (UID: \"dfa370e6-e90e-4a55-b8d0-332f370999b1\") " pod="openshift-must-gather-q6d7g/crc-debug-v896p" Oct 08 10:32:40 crc kubenswrapper[4744]: I1008 10:32:40.883580 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/crc-debug-v896p" Oct 08 10:32:41 crc kubenswrapper[4744]: I1008 10:32:41.148924 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q6d7g/crc-debug-v896p" event={"ID":"dfa370e6-e90e-4a55-b8d0-332f370999b1","Type":"ContainerStarted","Data":"eea22d3a75d984953e98ede236060ee94732dd7b1dea34d8888c1c10cdcff1da"} Oct 08 10:32:42 crc kubenswrapper[4744]: I1008 10:32:42.161809 4744 generic.go:334] "Generic (PLEG): container finished" podID="dfa370e6-e90e-4a55-b8d0-332f370999b1" containerID="d2468b2f263f6015723219012aab00a7f158f4ad6c0b1a82e526868be5133503" exitCode=0 Oct 08 10:32:42 crc kubenswrapper[4744]: I1008 10:32:42.162339 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q6d7g/crc-debug-v896p" event={"ID":"dfa370e6-e90e-4a55-b8d0-332f370999b1","Type":"ContainerDied","Data":"d2468b2f263f6015723219012aab00a7f158f4ad6c0b1a82e526868be5133503"} Oct 08 10:32:43 crc kubenswrapper[4744]: I1008 10:32:43.287262 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/crc-debug-v896p" Oct 08 10:32:43 crc kubenswrapper[4744]: I1008 10:32:43.432796 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwrdj\" (UniqueName: \"kubernetes.io/projected/dfa370e6-e90e-4a55-b8d0-332f370999b1-kube-api-access-zwrdj\") pod \"dfa370e6-e90e-4a55-b8d0-332f370999b1\" (UID: \"dfa370e6-e90e-4a55-b8d0-332f370999b1\") " Oct 08 10:32:43 crc kubenswrapper[4744]: I1008 10:32:43.432833 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dfa370e6-e90e-4a55-b8d0-332f370999b1-host\") pod \"dfa370e6-e90e-4a55-b8d0-332f370999b1\" (UID: \"dfa370e6-e90e-4a55-b8d0-332f370999b1\") " Oct 08 10:32:43 crc kubenswrapper[4744]: I1008 10:32:43.433752 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dfa370e6-e90e-4a55-b8d0-332f370999b1-host" (OuterVolumeSpecName: "host") pod "dfa370e6-e90e-4a55-b8d0-332f370999b1" (UID: "dfa370e6-e90e-4a55-b8d0-332f370999b1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 10:32:43 crc kubenswrapper[4744]: I1008 10:32:43.434462 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dfa370e6-e90e-4a55-b8d0-332f370999b1-host\") on node \"crc\" DevicePath \"\"" Oct 08 10:32:43 crc kubenswrapper[4744]: I1008 10:32:43.443457 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfa370e6-e90e-4a55-b8d0-332f370999b1-kube-api-access-zwrdj" (OuterVolumeSpecName: "kube-api-access-zwrdj") pod "dfa370e6-e90e-4a55-b8d0-332f370999b1" (UID: "dfa370e6-e90e-4a55-b8d0-332f370999b1"). InnerVolumeSpecName "kube-api-access-zwrdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:32:43 crc kubenswrapper[4744]: I1008 10:32:43.537113 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwrdj\" (UniqueName: \"kubernetes.io/projected/dfa370e6-e90e-4a55-b8d0-332f370999b1-kube-api-access-zwrdj\") on node \"crc\" DevicePath \"\"" Oct 08 10:32:44 crc kubenswrapper[4744]: I1008 10:32:44.210575 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q6d7g/crc-debug-v896p" event={"ID":"dfa370e6-e90e-4a55-b8d0-332f370999b1","Type":"ContainerDied","Data":"eea22d3a75d984953e98ede236060ee94732dd7b1dea34d8888c1c10cdcff1da"} Oct 08 10:32:44 crc kubenswrapper[4744]: I1008 10:32:44.210629 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eea22d3a75d984953e98ede236060ee94732dd7b1dea34d8888c1c10cdcff1da" Oct 08 10:32:44 crc kubenswrapper[4744]: I1008 10:32:44.210747 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/crc-debug-v896p" Oct 08 10:32:52 crc kubenswrapper[4744]: I1008 10:32:52.737420 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-q6d7g/crc-debug-v896p"] Oct 08 10:32:52 crc kubenswrapper[4744]: I1008 10:32:52.745883 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-q6d7g/crc-debug-v896p"] Oct 08 10:32:53 crc kubenswrapper[4744]: I1008 10:32:53.472240 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfa370e6-e90e-4a55-b8d0-332f370999b1" path="/var/lib/kubelet/pods/dfa370e6-e90e-4a55-b8d0-332f370999b1/volumes" Oct 08 10:32:54 crc kubenswrapper[4744]: I1008 10:32:54.072498 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-q6d7g/crc-debug-mtsp4"] Oct 08 10:32:54 crc kubenswrapper[4744]: E1008 10:32:54.073011 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfa370e6-e90e-4a55-b8d0-332f370999b1" containerName="container-00" Oct 08 10:32:54 crc kubenswrapper[4744]: I1008 10:32:54.073030 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfa370e6-e90e-4a55-b8d0-332f370999b1" containerName="container-00" Oct 08 10:32:54 crc kubenswrapper[4744]: I1008 10:32:54.073394 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfa370e6-e90e-4a55-b8d0-332f370999b1" containerName="container-00" Oct 08 10:32:54 crc kubenswrapper[4744]: I1008 10:32:54.074174 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/crc-debug-mtsp4" Oct 08 10:32:54 crc kubenswrapper[4744]: I1008 10:32:54.089112 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-q6d7g"/"default-dockercfg-88nh5" Oct 08 10:32:54 crc kubenswrapper[4744]: I1008 10:32:54.198572 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzsdk\" (UniqueName: \"kubernetes.io/projected/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b-kube-api-access-qzsdk\") pod \"crc-debug-mtsp4\" (UID: \"b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b\") " pod="openshift-must-gather-q6d7g/crc-debug-mtsp4" Oct 08 10:32:54 crc kubenswrapper[4744]: I1008 10:32:54.198717 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b-host\") pod \"crc-debug-mtsp4\" (UID: \"b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b\") " pod="openshift-must-gather-q6d7g/crc-debug-mtsp4" Oct 08 10:32:54 crc kubenswrapper[4744]: I1008 10:32:54.299835 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzsdk\" (UniqueName: \"kubernetes.io/projected/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b-kube-api-access-qzsdk\") pod \"crc-debug-mtsp4\" (UID: \"b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b\") " pod="openshift-must-gather-q6d7g/crc-debug-mtsp4" Oct 08 10:32:54 crc kubenswrapper[4744]: I1008 10:32:54.299958 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b-host\") pod \"crc-debug-mtsp4\" (UID: \"b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b\") " pod="openshift-must-gather-q6d7g/crc-debug-mtsp4" Oct 08 10:32:54 crc kubenswrapper[4744]: I1008 10:32:54.300055 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b-host\") pod \"crc-debug-mtsp4\" (UID: \"b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b\") " pod="openshift-must-gather-q6d7g/crc-debug-mtsp4" Oct 08 10:32:54 crc kubenswrapper[4744]: I1008 10:32:54.404741 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzsdk\" (UniqueName: \"kubernetes.io/projected/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b-kube-api-access-qzsdk\") pod \"crc-debug-mtsp4\" (UID: \"b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b\") " pod="openshift-must-gather-q6d7g/crc-debug-mtsp4" Oct 08 10:32:54 crc kubenswrapper[4744]: I1008 10:32:54.695308 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/crc-debug-mtsp4" Oct 08 10:32:55 crc kubenswrapper[4744]: I1008 10:32:55.343133 4744 generic.go:334] "Generic (PLEG): container finished" podID="b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b" containerID="6a84e929755ec01ea5211469b5f4eef0b9473f306e6eb59ecebc49fa844f635a" exitCode=0 Oct 08 10:32:55 crc kubenswrapper[4744]: I1008 10:32:55.343470 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q6d7g/crc-debug-mtsp4" event={"ID":"b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b","Type":"ContainerDied","Data":"6a84e929755ec01ea5211469b5f4eef0b9473f306e6eb59ecebc49fa844f635a"} Oct 08 10:32:55 crc kubenswrapper[4744]: I1008 10:32:55.343497 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q6d7g/crc-debug-mtsp4" event={"ID":"b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b","Type":"ContainerStarted","Data":"43c28826c9ebe68a43454538360c40e047cc962c282bf8af02be82b7fdb2826e"} Oct 08 10:32:55 crc kubenswrapper[4744]: I1008 10:32:55.379428 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-q6d7g/crc-debug-mtsp4"] Oct 08 10:32:55 crc kubenswrapper[4744]: I1008 10:32:55.384034 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-q6d7g/crc-debug-mtsp4"] Oct 08 10:32:55 crc kubenswrapper[4744]: I1008 10:32:55.460287 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_4c8e37bd-80bc-4e95-8baf-f571adecb232/memcached/0.log" Oct 08 10:32:56 crc kubenswrapper[4744]: I1008 10:32:56.475338 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/crc-debug-mtsp4" Oct 08 10:32:56 crc kubenswrapper[4744]: I1008 10:32:56.554010 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzsdk\" (UniqueName: \"kubernetes.io/projected/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b-kube-api-access-qzsdk\") pod \"b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b\" (UID: \"b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b\") " Oct 08 10:32:56 crc kubenswrapper[4744]: I1008 10:32:56.554137 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b-host\") pod \"b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b\" (UID: \"b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b\") " Oct 08 10:32:56 crc kubenswrapper[4744]: I1008 10:32:56.554237 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b-host" (OuterVolumeSpecName: "host") pod "b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b" (UID: "b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 10:32:56 crc kubenswrapper[4744]: I1008 10:32:56.555773 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b-host\") on node \"crc\" DevicePath \"\"" Oct 08 10:32:56 crc kubenswrapper[4744]: I1008 10:32:56.576556 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b-kube-api-access-qzsdk" (OuterVolumeSpecName: "kube-api-access-qzsdk") pod "b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b" (UID: "b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b"). InnerVolumeSpecName "kube-api-access-qzsdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:32:56 crc kubenswrapper[4744]: I1008 10:32:56.657829 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzsdk\" (UniqueName: \"kubernetes.io/projected/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b-kube-api-access-qzsdk\") on node \"crc\" DevicePath \"\"" Oct 08 10:32:57 crc kubenswrapper[4744]: I1008 10:32:57.363698 4744 scope.go:117] "RemoveContainer" containerID="6a84e929755ec01ea5211469b5f4eef0b9473f306e6eb59ecebc49fa844f635a" Oct 08 10:32:57 crc kubenswrapper[4744]: I1008 10:32:57.363767 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/crc-debug-mtsp4" Oct 08 10:32:57 crc kubenswrapper[4744]: I1008 10:32:57.477536 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b" path="/var/lib/kubelet/pods/b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b/volumes" Oct 08 10:33:11 crc kubenswrapper[4744]: I1008 10:33:11.647840 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-nzjx2_6d6f4234-2b91-4399-8aa2-f1064fde501c/kube-rbac-proxy/0.log" Oct 08 10:33:11 crc kubenswrapper[4744]: I1008 10:33:11.696456 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-nzjx2_6d6f4234-2b91-4399-8aa2-f1064fde501c/manager/0.log" Oct 08 10:33:11 crc kubenswrapper[4744]: I1008 10:33:11.910907 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-csxmb_8a43b240-9fe3-4c10-afa9-7e78777e76d5/kube-rbac-proxy/0.log" Oct 08 10:33:11 crc kubenswrapper[4744]: I1008 10:33:11.990783 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-csxmb_8a43b240-9fe3-4c10-afa9-7e78777e76d5/manager/0.log" Oct 08 10:33:12 crc kubenswrapper[4744]: I1008 10:33:12.173257 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-b9ntl_254b5ec4-bc16-4e8a-a61b-8a4a588dd629/kube-rbac-proxy/0.log" Oct 08 10:33:12 crc kubenswrapper[4744]: I1008 10:33:12.292862 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-b9ntl_254b5ec4-bc16-4e8a-a61b-8a4a588dd629/manager/0.log" Oct 08 10:33:12 crc kubenswrapper[4744]: I1008 10:33:12.449395 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/util/0.log" Oct 08 10:33:12 crc kubenswrapper[4744]: I1008 10:33:12.747448 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/util/0.log" Oct 08 10:33:12 crc kubenswrapper[4744]: I1008 10:33:12.796228 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/pull/0.log" Oct 08 10:33:12 crc kubenswrapper[4744]: I1008 10:33:12.810232 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/pull/0.log" Oct 08 10:33:13 crc kubenswrapper[4744]: I1008 10:33:13.041807 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/util/0.log" Oct 08 10:33:13 crc kubenswrapper[4744]: I1008 10:33:13.129236 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/pull/0.log" Oct 08 10:33:13 crc kubenswrapper[4744]: I1008 10:33:13.129597 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_eee899de00471dca260c21f15d574ae705f89c0888bdad088fc990c0ecgb6xw_db93bdf5-bdc7-44ba-b635-83ae935f947b/extract/0.log" Oct 08 10:33:13 crc kubenswrapper[4744]: I1008 10:33:13.340050 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-m9prk_622e796c-d598-4721-89bd-c0cb3f83a1ad/kube-rbac-proxy/0.log" Oct 08 10:33:13 crc kubenswrapper[4744]: I1008 10:33:13.382223 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-f9zmb_44cd0502-66fb-45ef-888a-da6273df7056/kube-rbac-proxy/0.log" Oct 08 10:33:13 crc kubenswrapper[4744]: I1008 10:33:13.487668 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-f9zmb_44cd0502-66fb-45ef-888a-da6273df7056/manager/0.log" Oct 08 10:33:14 crc kubenswrapper[4744]: I1008 10:33:14.209757 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-m9prk_622e796c-d598-4721-89bd-c0cb3f83a1ad/manager/0.log" Oct 08 10:33:14 crc kubenswrapper[4744]: I1008 10:33:14.295957 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-hsmrl_18a79d6e-9afe-43b5-8cf1-b5a991d2fd49/kube-rbac-proxy/0.log" Oct 08 10:33:14 crc kubenswrapper[4744]: I1008 10:33:14.316400 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-hsmrl_18a79d6e-9afe-43b5-8cf1-b5a991d2fd49/manager/0.log" Oct 08 10:33:14 crc kubenswrapper[4744]: I1008 10:33:14.586802 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-tbx59_b2bb8369-17dd-4f3b-8e8d-4af7895d892c/kube-rbac-proxy/0.log" Oct 08 10:33:14 crc kubenswrapper[4744]: I1008 10:33:14.741338 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-q9hcb_6126b107-8c51-4467-9c40-e2c4b268cec5/kube-rbac-proxy/0.log" Oct 08 10:33:14 crc kubenswrapper[4744]: I1008 10:33:14.936885 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-q9hcb_6126b107-8c51-4467-9c40-e2c4b268cec5/manager/0.log" Oct 08 10:33:15 crc kubenswrapper[4744]: I1008 10:33:15.082245 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-xjxx8_72a5635b-6920-4c2b-b91c-7621bc933959/kube-rbac-proxy/0.log" Oct 08 10:33:15 crc kubenswrapper[4744]: I1008 10:33:15.164138 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-tbx59_b2bb8369-17dd-4f3b-8e8d-4af7895d892c/manager/0.log" Oct 08 10:33:15 crc kubenswrapper[4744]: I1008 10:33:15.317175 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-lj9pr_58decb1a-987e-4366-9bf8-ad4bf73e5969/kube-rbac-proxy/0.log" Oct 08 10:33:15 crc kubenswrapper[4744]: I1008 10:33:15.335732 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-xjxx8_72a5635b-6920-4c2b-b91c-7621bc933959/manager/0.log" Oct 08 10:33:15 crc kubenswrapper[4744]: I1008 10:33:15.465195 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-lj9pr_58decb1a-987e-4366-9bf8-ad4bf73e5969/manager/0.log" Oct 08 10:33:16 crc kubenswrapper[4744]: I1008 10:33:16.093788 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-4xvzd_5fae63cb-c599-447b-bafb-2ed6bad836d3/kube-rbac-proxy/0.log" Oct 08 10:33:16 crc kubenswrapper[4744]: I1008 10:33:16.200021 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-4xvzd_5fae63cb-c599-447b-bafb-2ed6bad836d3/manager/0.log" Oct 08 10:33:16 crc kubenswrapper[4744]: I1008 10:33:16.236521 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-xl4hw_48158343-3842-44df-86c8-3a2f0e06a09c/kube-rbac-proxy/0.log" Oct 08 10:33:16 crc kubenswrapper[4744]: I1008 10:33:16.714941 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-xl4hw_48158343-3842-44df-86c8-3a2f0e06a09c/manager/0.log" Oct 08 10:33:16 crc kubenswrapper[4744]: I1008 10:33:16.891680 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-5vz2c_2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34/kube-rbac-proxy/0.log" Oct 08 10:33:16 crc kubenswrapper[4744]: I1008 10:33:16.985212 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-5vz2c_2f5a3cef-f2a2-4ddf-9deb-3438acfa6e34/manager/0.log" Oct 08 10:33:17 crc kubenswrapper[4744]: I1008 10:33:17.013566 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-crqx2_3eeb379f-9005-4908-ac9e-6dfb06405fb2/kube-rbac-proxy/0.log" Oct 08 10:33:17 crc kubenswrapper[4744]: I1008 10:33:17.033462 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-crqx2_3eeb379f-9005-4908-ac9e-6dfb06405fb2/manager/0.log" Oct 08 10:33:17 crc kubenswrapper[4744]: I1008 10:33:17.208443 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-747747dfccwc6fw_c5096473-ff21-474e-b7fc-03196e4ee5a8/kube-rbac-proxy/0.log" Oct 08 10:33:17 crc kubenswrapper[4744]: I1008 10:33:17.222407 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-747747dfccwc6fw_c5096473-ff21-474e-b7fc-03196e4ee5a8/manager/0.log" Oct 08 10:33:17 crc kubenswrapper[4744]: I1008 10:33:17.285470 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8bc6b8f5b-7q7gz_6fdbcdf9-cab4-4d8a-9270-2a121c83dc04/kube-rbac-proxy/0.log" Oct 08 10:33:17 crc kubenswrapper[4744]: I1008 10:33:17.617307 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-55f65988b-bjktr_b575f955-1e82-4afa-a84b-842c2ba7e47b/kube-rbac-proxy/0.log" Oct 08 10:33:17 crc kubenswrapper[4744]: I1008 10:33:17.861317 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-55f65988b-bjktr_b575f955-1e82-4afa-a84b-842c2ba7e47b/operator/0.log" Oct 08 10:33:17 crc kubenswrapper[4744]: I1008 10:33:17.913170 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79db49b9fb-qppfc_e991d608-9aa2-4028-8080-a6bd7dca66c8/kube-rbac-proxy/0.log" Oct 08 10:33:17 crc kubenswrapper[4744]: I1008 10:33:17.999208 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tc4hq_7f7736f3-f7de-4ff0-ae46-f01bf32693f5/registry-server/0.log" Oct 08 10:33:18 crc kubenswrapper[4744]: I1008 10:33:18.210242 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79db49b9fb-qppfc_e991d608-9aa2-4028-8080-a6bd7dca66c8/manager/0.log" Oct 08 10:33:18 crc kubenswrapper[4744]: I1008 10:33:18.330771 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-lbvtv_3a897464-5474-43db-99d5-35f691f64220/kube-rbac-proxy/0.log" Oct 08 10:33:18 crc kubenswrapper[4744]: I1008 10:33:18.335219 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-lbvtv_3a897464-5474-43db-99d5-35f691f64220/manager/0.log" Oct 08 10:33:18 crc kubenswrapper[4744]: I1008 10:33:18.519496 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8bc6b8f5b-7q7gz_6fdbcdf9-cab4-4d8a-9270-2a121c83dc04/manager/0.log" Oct 08 10:33:18 crc kubenswrapper[4744]: I1008 10:33:18.658274 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-frfr7_663cead5-7b5e-4940-94f9-5974f0424eda/kube-rbac-proxy/0.log" Oct 08 10:33:18 crc kubenswrapper[4744]: I1008 10:33:18.684115 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-ztld4_5bf79b31-d3bc-4422-9302-ac7a55a1d95d/operator/0.log" Oct 08 10:33:18 crc kubenswrapper[4744]: I1008 10:33:18.751725 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-frfr7_663cead5-7b5e-4940-94f9-5974f0424eda/manager/0.log" Oct 08 10:33:18 crc kubenswrapper[4744]: I1008 10:33:18.852970 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76796d4c6b-gv4tw_f3456365-6f0d-4087-b753-40292c80bf12/kube-rbac-proxy/0.log" Oct 08 10:33:19 crc kubenswrapper[4744]: I1008 10:33:19.003891 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76796d4c6b-gv4tw_f3456365-6f0d-4087-b753-40292c80bf12/manager/0.log" Oct 08 10:33:19 crc kubenswrapper[4744]: I1008 10:33:19.022435 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56c698c775-bf9cm_12b5913f-74df-4ba6-8aa8-af5989327ddd/kube-rbac-proxy/0.log" Oct 08 10:33:19 crc kubenswrapper[4744]: I1008 10:33:19.068812 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56c698c775-bf9cm_12b5913f-74df-4ba6-8aa8-af5989327ddd/manager/0.log" Oct 08 10:33:19 crc kubenswrapper[4744]: I1008 10:33:19.234646 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7794bc6bd-q884x_60f0c455-812d-40fe-b1b6-5aa75c34753c/manager/0.log" Oct 08 10:33:19 crc kubenswrapper[4744]: I1008 10:33:19.245485 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7794bc6bd-q884x_60f0c455-812d-40fe-b1b6-5aa75c34753c/kube-rbac-proxy/0.log" Oct 08 10:33:19 crc kubenswrapper[4744]: I1008 10:33:19.690818 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:33:19 crc kubenswrapper[4744]: I1008 10:33:19.690870 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:33:37 crc kubenswrapper[4744]: I1008 10:33:37.654958 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5nnfc_81cf0144-e970-4fe2-8663-15b006dc1b91/control-plane-machine-set-operator/0.log" Oct 08 10:33:37 crc kubenswrapper[4744]: I1008 10:33:37.675588 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-b5q8z_a7221db0-3f54-4e07-95ec-56c4fb95aecc/kube-rbac-proxy/0.log" Oct 08 10:33:37 crc kubenswrapper[4744]: I1008 10:33:37.859149 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-b5q8z_a7221db0-3f54-4e07-95ec-56c4fb95aecc/machine-api-operator/0.log" Oct 08 10:33:49 crc kubenswrapper[4744]: I1008 10:33:49.690118 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:33:49 crc kubenswrapper[4744]: I1008 10:33:49.690649 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:33:50 crc kubenswrapper[4744]: I1008 10:33:50.095111 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-87br5_5fde9304-c103-4e8a-96bc-d47a0dea1635/cert-manager-controller/0.log" Oct 08 10:33:50 crc kubenswrapper[4744]: I1008 10:33:50.386037 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-j5pzr_b2fe7e00-2fdb-4186-81db-c211ef88b772/cert-manager-cainjector/0.log" Oct 08 10:33:50 crc kubenswrapper[4744]: I1008 10:33:50.479517 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-4478x_39d1521a-c337-4259-bc31-0bf6357ce325/cert-manager-webhook/0.log" Oct 08 10:34:02 crc kubenswrapper[4744]: I1008 10:34:02.052470 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-hrjff_2c9804b5-0729-4059-bb00-35fd1d355b3c/nmstate-console-plugin/0.log" Oct 08 10:34:02 crc kubenswrapper[4744]: I1008 10:34:02.209017 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-tk2k5_8b5fc549-4718-4e89-82a9-b39e93a2220d/nmstate-handler/0.log" Oct 08 10:34:02 crc kubenswrapper[4744]: I1008 10:34:02.288303 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-9qfcx_db58dfb9-4fd9-4896-add0-116a1ed9079d/kube-rbac-proxy/0.log" Oct 08 10:34:02 crc kubenswrapper[4744]: I1008 10:34:02.329803 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-9qfcx_db58dfb9-4fd9-4896-add0-116a1ed9079d/nmstate-metrics/0.log" Oct 08 10:34:02 crc kubenswrapper[4744]: I1008 10:34:02.548663 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-nmnhx_1f1d4ca0-878e-4211-a40a-37f16085fe47/nmstate-operator/0.log" Oct 08 10:34:02 crc kubenswrapper[4744]: I1008 10:34:02.579649 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-pnz4l_89b615a3-0a5d-483c-8d41-2dee267f1686/nmstate-webhook/0.log" Oct 08 10:34:18 crc kubenswrapper[4744]: I1008 10:34:18.753361 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-8rkml_4f0fc93a-5521-4a7a-90d7-7b1794dc728f/kube-rbac-proxy/0.log" Oct 08 10:34:18 crc kubenswrapper[4744]: I1008 10:34:18.773648 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-8rkml_4f0fc93a-5521-4a7a-90d7-7b1794dc728f/controller/0.log" Oct 08 10:34:18 crc kubenswrapper[4744]: I1008 10:34:18.928080 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-frr-files/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.110197 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-reloader/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.113862 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-reloader/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.138519 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-metrics/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.189666 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-frr-files/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.377305 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-metrics/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.416914 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-metrics/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.417587 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-frr-files/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.469417 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-reloader/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.645356 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-frr-files/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.673812 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-metrics/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.686222 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/cp-reloader/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.691490 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.691540 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.691578 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.691988 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"54379f58e3765719921fb408b4441ea4d56d9cecce1a289805e1d0aad427c6de"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.692049 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://54379f58e3765719921fb408b4441ea4d56d9cecce1a289805e1d0aad427c6de" gracePeriod=600 Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.790669 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/controller/0.log" Oct 08 10:34:19 crc kubenswrapper[4744]: I1008 10:34:19.915860 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/frr-metrics/0.log" Oct 08 10:34:20 crc kubenswrapper[4744]: I1008 10:34:20.071978 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/kube-rbac-proxy/0.log" Oct 08 10:34:20 crc kubenswrapper[4744]: I1008 10:34:20.087792 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="54379f58e3765719921fb408b4441ea4d56d9cecce1a289805e1d0aad427c6de" exitCode=0 Oct 08 10:34:20 crc kubenswrapper[4744]: I1008 10:34:20.088125 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"54379f58e3765719921fb408b4441ea4d56d9cecce1a289805e1d0aad427c6de"} Oct 08 10:34:20 crc kubenswrapper[4744]: I1008 10:34:20.088166 4744 scope.go:117] "RemoveContainer" containerID="40ef35c0e44ec0461087a80142a1e5b443241e2a817fc652ea56300dc9b13dab" Oct 08 10:34:20 crc kubenswrapper[4744]: I1008 10:34:20.110886 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/kube-rbac-proxy-frr/0.log" Oct 08 10:34:20 crc kubenswrapper[4744]: I1008 10:34:20.443755 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-ptt9h_cf7d04b2-1136-4427-8939-e2974337725c/frr-k8s-webhook-server/0.log" Oct 08 10:34:20 crc kubenswrapper[4744]: I1008 10:34:20.447304 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/reloader/0.log" Oct 08 10:34:20 crc kubenswrapper[4744]: I1008 10:34:20.749278 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5998774b9-gk4vj_fb79b0ea-df0c-4f82-a297-4ce347bdc363/manager/0.log" Oct 08 10:34:21 crc kubenswrapper[4744]: I1008 10:34:21.072921 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-dbd787fb9-lsv2s_9de92423-8f60-43bd-842c-b47f51c9e20c/webhook-server/0.log" Oct 08 10:34:21 crc kubenswrapper[4744]: I1008 10:34:21.103714 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerStarted","Data":"8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685"} Oct 08 10:34:21 crc kubenswrapper[4744]: I1008 10:34:21.306084 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-89qxn_349b3e3f-87d0-4e3d-b2dc-c96452a7199c/kube-rbac-proxy/0.log" Oct 08 10:34:21 crc kubenswrapper[4744]: I1008 10:34:21.655829 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5skh4_38b9b97b-50ef-48c1-b917-a6c2f2be304b/frr/0.log" Oct 08 10:34:21 crc kubenswrapper[4744]: I1008 10:34:21.861095 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-89qxn_349b3e3f-87d0-4e3d-b2dc-c96452a7199c/speaker/0.log" Oct 08 10:34:35 crc kubenswrapper[4744]: I1008 10:34:35.314589 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/util/0.log" Oct 08 10:34:35 crc kubenswrapper[4744]: I1008 10:34:35.434782 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/util/0.log" Oct 08 10:34:35 crc kubenswrapper[4744]: I1008 10:34:35.471217 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/pull/0.log" Oct 08 10:34:35 crc kubenswrapper[4744]: I1008 10:34:35.493807 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/pull/0.log" Oct 08 10:34:35 crc kubenswrapper[4744]: I1008 10:34:35.682022 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/util/0.log" Oct 08 10:34:35 crc kubenswrapper[4744]: I1008 10:34:35.686903 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/extract/0.log" Oct 08 10:34:35 crc kubenswrapper[4744]: I1008 10:34:35.703915 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2tp9xp_efc6b240-277b-494e-9189-afd42c5a2aba/pull/0.log" Oct 08 10:34:35 crc kubenswrapper[4744]: I1008 10:34:35.878884 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/extract-utilities/0.log" Oct 08 10:34:36 crc kubenswrapper[4744]: I1008 10:34:36.552698 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/extract-content/0.log" Oct 08 10:34:36 crc kubenswrapper[4744]: I1008 10:34:36.557667 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/extract-utilities/0.log" Oct 08 10:34:36 crc kubenswrapper[4744]: I1008 10:34:36.567135 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/extract-content/0.log" Oct 08 10:34:36 crc kubenswrapper[4744]: I1008 10:34:36.712326 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/extract-utilities/0.log" Oct 08 10:34:36 crc kubenswrapper[4744]: I1008 10:34:36.766647 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/extract-content/0.log" Oct 08 10:34:37 crc kubenswrapper[4744]: I1008 10:34:37.064910 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/extract-utilities/0.log" Oct 08 10:34:37 crc kubenswrapper[4744]: I1008 10:34:37.314926 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/extract-utilities/0.log" Oct 08 10:34:37 crc kubenswrapper[4744]: I1008 10:34:37.342928 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5gqrp_003a8292-d26b-47f8-89ce-60165fcad99e/registry-server/0.log" Oct 08 10:34:37 crc kubenswrapper[4744]: I1008 10:34:37.361312 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/extract-content/0.log" Oct 08 10:34:37 crc kubenswrapper[4744]: I1008 10:34:37.434987 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/extract-content/0.log" Oct 08 10:34:37 crc kubenswrapper[4744]: I1008 10:34:37.676892 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/extract-content/0.log" Oct 08 10:34:37 crc kubenswrapper[4744]: I1008 10:34:37.681651 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/extract-utilities/0.log" Oct 08 10:34:38 crc kubenswrapper[4744]: I1008 10:34:38.007462 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/util/0.log" Oct 08 10:34:38 crc kubenswrapper[4744]: I1008 10:34:38.266483 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ffptv_12c528d4-e939-4c33-a145-cf8e58bb8c89/registry-server/0.log" Oct 08 10:34:38 crc kubenswrapper[4744]: I1008 10:34:38.319692 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/pull/0.log" Oct 08 10:34:38 crc kubenswrapper[4744]: I1008 10:34:38.335989 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/pull/0.log" Oct 08 10:34:38 crc kubenswrapper[4744]: I1008 10:34:38.361357 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/util/0.log" Oct 08 10:34:38 crc kubenswrapper[4744]: I1008 10:34:38.509934 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/util/0.log" Oct 08 10:34:38 crc kubenswrapper[4744]: I1008 10:34:38.562970 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/pull/0.log" Oct 08 10:34:38 crc kubenswrapper[4744]: I1008 10:34:38.603447 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835clwk65_8a3e5f3c-8790-47bb-a646-97dbbf595559/extract/0.log" Oct 08 10:34:38 crc kubenswrapper[4744]: I1008 10:34:38.746777 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-rjzwz_0364f287-922c-410f-9cee-1866cdba42e1/marketplace-operator/0.log" Oct 08 10:34:38 crc kubenswrapper[4744]: I1008 10:34:38.834415 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/extract-utilities/0.log" Oct 08 10:34:38 crc kubenswrapper[4744]: I1008 10:34:38.947312 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/extract-utilities/0.log" Oct 08 10:34:38 crc kubenswrapper[4744]: I1008 10:34:38.960520 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/extract-content/0.log" Oct 08 10:34:39 crc kubenswrapper[4744]: I1008 10:34:39.001424 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/extract-content/0.log" Oct 08 10:34:39 crc kubenswrapper[4744]: I1008 10:34:39.157970 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/extract-utilities/0.log" Oct 08 10:34:39 crc kubenswrapper[4744]: I1008 10:34:39.164610 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/extract-content/0.log" Oct 08 10:34:39 crc kubenswrapper[4744]: I1008 10:34:39.255838 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/extract-utilities/0.log" Oct 08 10:34:39 crc kubenswrapper[4744]: I1008 10:34:39.414459 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8s9r_8c8a122a-bcbf-474b-8c68-ed54aec6a6f4/registry-server/0.log" Oct 08 10:34:39 crc kubenswrapper[4744]: I1008 10:34:39.480484 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/extract-utilities/0.log" Oct 08 10:34:39 crc kubenswrapper[4744]: I1008 10:34:39.521335 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/extract-content/0.log" Oct 08 10:34:39 crc kubenswrapper[4744]: I1008 10:34:39.557918 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/extract-content/0.log" Oct 08 10:34:39 crc kubenswrapper[4744]: I1008 10:34:39.843315 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/extract-content/0.log" Oct 08 10:34:39 crc kubenswrapper[4744]: I1008 10:34:39.860136 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/extract-utilities/0.log" Oct 08 10:34:40 crc kubenswrapper[4744]: I1008 10:34:40.418614 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-q9562_41990ef2-32ac-4477-a702-19e6fd228546/registry-server/0.log" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.488270 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l9z9p"] Oct 08 10:36:18 crc kubenswrapper[4744]: E1008 10:36:18.489244 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b" containerName="container-00" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.489270 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b" containerName="container-00" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.489547 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8d5ca51-2fe4-4781-9d5e-5ea430f9d15b" containerName="container-00" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.490934 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.504316 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l9z9p"] Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.548849 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d979fefe-1861-4010-85b7-b709af321101-catalog-content\") pod \"community-operators-l9z9p\" (UID: \"d979fefe-1861-4010-85b7-b709af321101\") " pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.549013 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnlnc\" (UniqueName: \"kubernetes.io/projected/d979fefe-1861-4010-85b7-b709af321101-kube-api-access-rnlnc\") pod \"community-operators-l9z9p\" (UID: \"d979fefe-1861-4010-85b7-b709af321101\") " pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.549078 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d979fefe-1861-4010-85b7-b709af321101-utilities\") pod \"community-operators-l9z9p\" (UID: \"d979fefe-1861-4010-85b7-b709af321101\") " pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.650273 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnlnc\" (UniqueName: \"kubernetes.io/projected/d979fefe-1861-4010-85b7-b709af321101-kube-api-access-rnlnc\") pod \"community-operators-l9z9p\" (UID: \"d979fefe-1861-4010-85b7-b709af321101\") " pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.650335 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d979fefe-1861-4010-85b7-b709af321101-utilities\") pod \"community-operators-l9z9p\" (UID: \"d979fefe-1861-4010-85b7-b709af321101\") " pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.650425 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d979fefe-1861-4010-85b7-b709af321101-catalog-content\") pod \"community-operators-l9z9p\" (UID: \"d979fefe-1861-4010-85b7-b709af321101\") " pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.651042 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d979fefe-1861-4010-85b7-b709af321101-catalog-content\") pod \"community-operators-l9z9p\" (UID: \"d979fefe-1861-4010-85b7-b709af321101\") " pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.651227 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d979fefe-1861-4010-85b7-b709af321101-utilities\") pod \"community-operators-l9z9p\" (UID: \"d979fefe-1861-4010-85b7-b709af321101\") " pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.673993 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnlnc\" (UniqueName: \"kubernetes.io/projected/d979fefe-1861-4010-85b7-b709af321101-kube-api-access-rnlnc\") pod \"community-operators-l9z9p\" (UID: \"d979fefe-1861-4010-85b7-b709af321101\") " pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:18 crc kubenswrapper[4744]: I1008 10:36:18.823059 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:19 crc kubenswrapper[4744]: I1008 10:36:19.540755 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l9z9p"] Oct 08 10:36:20 crc kubenswrapper[4744]: I1008 10:36:20.178425 4744 generic.go:334] "Generic (PLEG): container finished" podID="d979fefe-1861-4010-85b7-b709af321101" containerID="6a03baf10280b639e321eb1b9ecf4313e4a7aa81c381425524bad84305a04f93" exitCode=0 Oct 08 10:36:20 crc kubenswrapper[4744]: I1008 10:36:20.178511 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9z9p" event={"ID":"d979fefe-1861-4010-85b7-b709af321101","Type":"ContainerDied","Data":"6a03baf10280b639e321eb1b9ecf4313e4a7aa81c381425524bad84305a04f93"} Oct 08 10:36:20 crc kubenswrapper[4744]: I1008 10:36:20.178726 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9z9p" event={"ID":"d979fefe-1861-4010-85b7-b709af321101","Type":"ContainerStarted","Data":"eda7a61c7da65e14edd815fde8f238f4afde51b025e3d5107b6f5a7bf0c6b093"} Oct 08 10:36:20 crc kubenswrapper[4744]: I1008 10:36:20.181679 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 10:36:22 crc kubenswrapper[4744]: I1008 10:36:22.209456 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9z9p" event={"ID":"d979fefe-1861-4010-85b7-b709af321101","Type":"ContainerStarted","Data":"775601d4e62568b39698bee8336bc3d89d0395109fa7797df69f6dc24716990b"} Oct 08 10:36:24 crc kubenswrapper[4744]: I1008 10:36:24.230425 4744 generic.go:334] "Generic (PLEG): container finished" podID="d979fefe-1861-4010-85b7-b709af321101" containerID="775601d4e62568b39698bee8336bc3d89d0395109fa7797df69f6dc24716990b" exitCode=0 Oct 08 10:36:24 crc kubenswrapper[4744]: I1008 10:36:24.230463 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9z9p" event={"ID":"d979fefe-1861-4010-85b7-b709af321101","Type":"ContainerDied","Data":"775601d4e62568b39698bee8336bc3d89d0395109fa7797df69f6dc24716990b"} Oct 08 10:36:25 crc kubenswrapper[4744]: I1008 10:36:25.245821 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9z9p" event={"ID":"d979fefe-1861-4010-85b7-b709af321101","Type":"ContainerStarted","Data":"25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647"} Oct 08 10:36:25 crc kubenswrapper[4744]: I1008 10:36:25.300280 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l9z9p" podStartSLOduration=2.823926342 podStartE2EDuration="7.300258339s" podCreationTimestamp="2025-10-08 10:36:18 +0000 UTC" firstStartedPulling="2025-10-08 10:36:20.181410084 +0000 UTC m=+5075.429055323" lastFinishedPulling="2025-10-08 10:36:24.657742081 +0000 UTC m=+5079.905387320" observedRunningTime="2025-10-08 10:36:25.296988868 +0000 UTC m=+5080.544634117" watchObservedRunningTime="2025-10-08 10:36:25.300258339 +0000 UTC m=+5080.547903568" Oct 08 10:36:28 crc kubenswrapper[4744]: I1008 10:36:28.824023 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:28 crc kubenswrapper[4744]: I1008 10:36:28.825105 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:28 crc kubenswrapper[4744]: I1008 10:36:28.871938 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:29 crc kubenswrapper[4744]: I1008 10:36:29.319457 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:30 crc kubenswrapper[4744]: I1008 10:36:30.107552 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l9z9p"] Oct 08 10:36:31 crc kubenswrapper[4744]: I1008 10:36:31.296933 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l9z9p" podUID="d979fefe-1861-4010-85b7-b709af321101" containerName="registry-server" containerID="cri-o://25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647" gracePeriod=2 Oct 08 10:36:31 crc kubenswrapper[4744]: I1008 10:36:31.730943 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:31 crc kubenswrapper[4744]: I1008 10:36:31.830580 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d979fefe-1861-4010-85b7-b709af321101-catalog-content\") pod \"d979fefe-1861-4010-85b7-b709af321101\" (UID: \"d979fefe-1861-4010-85b7-b709af321101\") " Oct 08 10:36:31 crc kubenswrapper[4744]: I1008 10:36:31.830769 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnlnc\" (UniqueName: \"kubernetes.io/projected/d979fefe-1861-4010-85b7-b709af321101-kube-api-access-rnlnc\") pod \"d979fefe-1861-4010-85b7-b709af321101\" (UID: \"d979fefe-1861-4010-85b7-b709af321101\") " Oct 08 10:36:31 crc kubenswrapper[4744]: I1008 10:36:31.830848 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d979fefe-1861-4010-85b7-b709af321101-utilities\") pod \"d979fefe-1861-4010-85b7-b709af321101\" (UID: \"d979fefe-1861-4010-85b7-b709af321101\") " Oct 08 10:36:31 crc kubenswrapper[4744]: I1008 10:36:31.832269 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d979fefe-1861-4010-85b7-b709af321101-utilities" (OuterVolumeSpecName: "utilities") pod "d979fefe-1861-4010-85b7-b709af321101" (UID: "d979fefe-1861-4010-85b7-b709af321101"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:36:31 crc kubenswrapper[4744]: I1008 10:36:31.843538 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d979fefe-1861-4010-85b7-b709af321101-kube-api-access-rnlnc" (OuterVolumeSpecName: "kube-api-access-rnlnc") pod "d979fefe-1861-4010-85b7-b709af321101" (UID: "d979fefe-1861-4010-85b7-b709af321101"). InnerVolumeSpecName "kube-api-access-rnlnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:36:31 crc kubenswrapper[4744]: I1008 10:36:31.880098 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d979fefe-1861-4010-85b7-b709af321101-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d979fefe-1861-4010-85b7-b709af321101" (UID: "d979fefe-1861-4010-85b7-b709af321101"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:36:31 crc kubenswrapper[4744]: I1008 10:36:31.933068 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d979fefe-1861-4010-85b7-b709af321101-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:36:31 crc kubenswrapper[4744]: I1008 10:36:31.933308 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnlnc\" (UniqueName: \"kubernetes.io/projected/d979fefe-1861-4010-85b7-b709af321101-kube-api-access-rnlnc\") on node \"crc\" DevicePath \"\"" Oct 08 10:36:31 crc kubenswrapper[4744]: I1008 10:36:31.933467 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d979fefe-1861-4010-85b7-b709af321101-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.307601 4744 generic.go:334] "Generic (PLEG): container finished" podID="d979fefe-1861-4010-85b7-b709af321101" containerID="25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647" exitCode=0 Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.307901 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l9z9p" Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.307820 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9z9p" event={"ID":"d979fefe-1861-4010-85b7-b709af321101","Type":"ContainerDied","Data":"25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647"} Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.309096 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l9z9p" event={"ID":"d979fefe-1861-4010-85b7-b709af321101","Type":"ContainerDied","Data":"eda7a61c7da65e14edd815fde8f238f4afde51b025e3d5107b6f5a7bf0c6b093"} Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.309120 4744 scope.go:117] "RemoveContainer" containerID="25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647" Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.343235 4744 scope.go:117] "RemoveContainer" containerID="775601d4e62568b39698bee8336bc3d89d0395109fa7797df69f6dc24716990b" Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.346138 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l9z9p"] Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.353237 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l9z9p"] Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.826107 4744 scope.go:117] "RemoveContainer" containerID="6a03baf10280b639e321eb1b9ecf4313e4a7aa81c381425524bad84305a04f93" Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.863579 4744 scope.go:117] "RemoveContainer" containerID="25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647" Oct 08 10:36:32 crc kubenswrapper[4744]: E1008 10:36:32.863976 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647\": container with ID starting with 25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647 not found: ID does not exist" containerID="25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647" Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.864007 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647"} err="failed to get container status \"25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647\": rpc error: code = NotFound desc = could not find container \"25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647\": container with ID starting with 25ba17f705ebcf920e9507d1b4160014e0625bf0c36b285defcfea907d064647 not found: ID does not exist" Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.864028 4744 scope.go:117] "RemoveContainer" containerID="775601d4e62568b39698bee8336bc3d89d0395109fa7797df69f6dc24716990b" Oct 08 10:36:32 crc kubenswrapper[4744]: E1008 10:36:32.864262 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"775601d4e62568b39698bee8336bc3d89d0395109fa7797df69f6dc24716990b\": container with ID starting with 775601d4e62568b39698bee8336bc3d89d0395109fa7797df69f6dc24716990b not found: ID does not exist" containerID="775601d4e62568b39698bee8336bc3d89d0395109fa7797df69f6dc24716990b" Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.864286 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"775601d4e62568b39698bee8336bc3d89d0395109fa7797df69f6dc24716990b"} err="failed to get container status \"775601d4e62568b39698bee8336bc3d89d0395109fa7797df69f6dc24716990b\": rpc error: code = NotFound desc = could not find container \"775601d4e62568b39698bee8336bc3d89d0395109fa7797df69f6dc24716990b\": container with ID starting with 775601d4e62568b39698bee8336bc3d89d0395109fa7797df69f6dc24716990b not found: ID does not exist" Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.864300 4744 scope.go:117] "RemoveContainer" containerID="6a03baf10280b639e321eb1b9ecf4313e4a7aa81c381425524bad84305a04f93" Oct 08 10:36:32 crc kubenswrapper[4744]: E1008 10:36:32.864548 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a03baf10280b639e321eb1b9ecf4313e4a7aa81c381425524bad84305a04f93\": container with ID starting with 6a03baf10280b639e321eb1b9ecf4313e4a7aa81c381425524bad84305a04f93 not found: ID does not exist" containerID="6a03baf10280b639e321eb1b9ecf4313e4a7aa81c381425524bad84305a04f93" Oct 08 10:36:32 crc kubenswrapper[4744]: I1008 10:36:32.864573 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a03baf10280b639e321eb1b9ecf4313e4a7aa81c381425524bad84305a04f93"} err="failed to get container status \"6a03baf10280b639e321eb1b9ecf4313e4a7aa81c381425524bad84305a04f93\": rpc error: code = NotFound desc = could not find container \"6a03baf10280b639e321eb1b9ecf4313e4a7aa81c381425524bad84305a04f93\": container with ID starting with 6a03baf10280b639e321eb1b9ecf4313e4a7aa81c381425524bad84305a04f93 not found: ID does not exist" Oct 08 10:36:33 crc kubenswrapper[4744]: I1008 10:36:33.483330 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d979fefe-1861-4010-85b7-b709af321101" path="/var/lib/kubelet/pods/d979fefe-1861-4010-85b7-b709af321101/volumes" Oct 08 10:36:49 crc kubenswrapper[4744]: I1008 10:36:49.690144 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:36:49 crc kubenswrapper[4744]: I1008 10:36:49.690863 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:37:19 crc kubenswrapper[4744]: I1008 10:37:19.690277 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:37:19 crc kubenswrapper[4744]: I1008 10:37:19.690819 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:37:25 crc kubenswrapper[4744]: I1008 10:37:25.878100 4744 generic.go:334] "Generic (PLEG): container finished" podID="de13d6e4-3524-4157-837a-eec911bc5316" containerID="27b64049af580689f69aab067417d368043c7d29952c292ba8476841a92abace" exitCode=0 Oct 08 10:37:25 crc kubenswrapper[4744]: I1008 10:37:25.878177 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q6d7g/must-gather-f52xr" event={"ID":"de13d6e4-3524-4157-837a-eec911bc5316","Type":"ContainerDied","Data":"27b64049af580689f69aab067417d368043c7d29952c292ba8476841a92abace"} Oct 08 10:37:25 crc kubenswrapper[4744]: I1008 10:37:25.879063 4744 scope.go:117] "RemoveContainer" containerID="27b64049af580689f69aab067417d368043c7d29952c292ba8476841a92abace" Oct 08 10:37:26 crc kubenswrapper[4744]: I1008 10:37:26.622566 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-q6d7g_must-gather-f52xr_de13d6e4-3524-4157-837a-eec911bc5316/gather/0.log" Oct 08 10:37:41 crc kubenswrapper[4744]: I1008 10:37:41.687230 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-q6d7g/must-gather-f52xr"] Oct 08 10:37:41 crc kubenswrapper[4744]: I1008 10:37:41.687922 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-q6d7g/must-gather-f52xr" podUID="de13d6e4-3524-4157-837a-eec911bc5316" containerName="copy" containerID="cri-o://54cb51c42a803df0b907d6d7cec344c30b0ca2884516490478d601c9fbbb6b0e" gracePeriod=2 Oct 08 10:37:41 crc kubenswrapper[4744]: I1008 10:37:41.701059 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-q6d7g/must-gather-f52xr"] Oct 08 10:37:42 crc kubenswrapper[4744]: I1008 10:37:42.045543 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-q6d7g_must-gather-f52xr_de13d6e4-3524-4157-837a-eec911bc5316/copy/0.log" Oct 08 10:37:42 crc kubenswrapper[4744]: I1008 10:37:42.046325 4744 generic.go:334] "Generic (PLEG): container finished" podID="de13d6e4-3524-4157-837a-eec911bc5316" containerID="54cb51c42a803df0b907d6d7cec344c30b0ca2884516490478d601c9fbbb6b0e" exitCode=143 Oct 08 10:37:42 crc kubenswrapper[4744]: I1008 10:37:42.324984 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-q6d7g_must-gather-f52xr_de13d6e4-3524-4157-837a-eec911bc5316/copy/0.log" Oct 08 10:37:42 crc kubenswrapper[4744]: I1008 10:37:42.327955 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/must-gather-f52xr" Oct 08 10:37:42 crc kubenswrapper[4744]: I1008 10:37:42.502335 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djm7m\" (UniqueName: \"kubernetes.io/projected/de13d6e4-3524-4157-837a-eec911bc5316-kube-api-access-djm7m\") pod \"de13d6e4-3524-4157-837a-eec911bc5316\" (UID: \"de13d6e4-3524-4157-837a-eec911bc5316\") " Oct 08 10:37:42 crc kubenswrapper[4744]: I1008 10:37:42.502538 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/de13d6e4-3524-4157-837a-eec911bc5316-must-gather-output\") pod \"de13d6e4-3524-4157-837a-eec911bc5316\" (UID: \"de13d6e4-3524-4157-837a-eec911bc5316\") " Oct 08 10:37:42 crc kubenswrapper[4744]: I1008 10:37:42.517263 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de13d6e4-3524-4157-837a-eec911bc5316-kube-api-access-djm7m" (OuterVolumeSpecName: "kube-api-access-djm7m") pod "de13d6e4-3524-4157-837a-eec911bc5316" (UID: "de13d6e4-3524-4157-837a-eec911bc5316"). InnerVolumeSpecName "kube-api-access-djm7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:37:42 crc kubenswrapper[4744]: I1008 10:37:42.605314 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djm7m\" (UniqueName: \"kubernetes.io/projected/de13d6e4-3524-4157-837a-eec911bc5316-kube-api-access-djm7m\") on node \"crc\" DevicePath \"\"" Oct 08 10:37:42 crc kubenswrapper[4744]: I1008 10:37:42.674142 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de13d6e4-3524-4157-837a-eec911bc5316-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "de13d6e4-3524-4157-837a-eec911bc5316" (UID: "de13d6e4-3524-4157-837a-eec911bc5316"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:37:42 crc kubenswrapper[4744]: I1008 10:37:42.706634 4744 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/de13d6e4-3524-4157-837a-eec911bc5316-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 08 10:37:43 crc kubenswrapper[4744]: I1008 10:37:43.055751 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-q6d7g_must-gather-f52xr_de13d6e4-3524-4157-837a-eec911bc5316/copy/0.log" Oct 08 10:37:43 crc kubenswrapper[4744]: I1008 10:37:43.056479 4744 scope.go:117] "RemoveContainer" containerID="54cb51c42a803df0b907d6d7cec344c30b0ca2884516490478d601c9fbbb6b0e" Oct 08 10:37:43 crc kubenswrapper[4744]: I1008 10:37:43.056503 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q6d7g/must-gather-f52xr" Oct 08 10:37:43 crc kubenswrapper[4744]: I1008 10:37:43.083587 4744 scope.go:117] "RemoveContainer" containerID="27b64049af580689f69aab067417d368043c7d29952c292ba8476841a92abace" Oct 08 10:37:43 crc kubenswrapper[4744]: I1008 10:37:43.462852 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de13d6e4-3524-4157-837a-eec911bc5316" path="/var/lib/kubelet/pods/de13d6e4-3524-4157-837a-eec911bc5316/volumes" Oct 08 10:37:49 crc kubenswrapper[4744]: I1008 10:37:49.690791 4744 patch_prober.go:28] interesting pod/machine-config-daemon-ttxzq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 10:37:49 crc kubenswrapper[4744]: I1008 10:37:49.691317 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 10:37:49 crc kubenswrapper[4744]: I1008 10:37:49.691357 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" Oct 08 10:37:49 crc kubenswrapper[4744]: I1008 10:37:49.692110 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685"} pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 10:37:49 crc kubenswrapper[4744]: I1008 10:37:49.692160 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerName="machine-config-daemon" containerID="cri-o://8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685" gracePeriod=600 Oct 08 10:37:49 crc kubenswrapper[4744]: E1008 10:37:49.815100 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:37:50 crc kubenswrapper[4744]: I1008 10:37:50.115162 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" containerID="8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685" exitCode=0 Oct 08 10:37:50 crc kubenswrapper[4744]: I1008 10:37:50.115191 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" event={"ID":"9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb","Type":"ContainerDied","Data":"8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685"} Oct 08 10:37:50 crc kubenswrapper[4744]: I1008 10:37:50.115404 4744 scope.go:117] "RemoveContainer" containerID="54379f58e3765719921fb408b4441ea4d56d9cecce1a289805e1d0aad427c6de" Oct 08 10:37:50 crc kubenswrapper[4744]: I1008 10:37:50.116716 4744 scope.go:117] "RemoveContainer" containerID="8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685" Oct 08 10:37:50 crc kubenswrapper[4744]: E1008 10:37:50.117067 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:38:01 crc kubenswrapper[4744]: I1008 10:38:01.453649 4744 scope.go:117] "RemoveContainer" containerID="8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685" Oct 08 10:38:01 crc kubenswrapper[4744]: E1008 10:38:01.455148 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:38:14 crc kubenswrapper[4744]: I1008 10:38:14.452869 4744 scope.go:117] "RemoveContainer" containerID="8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685" Oct 08 10:38:14 crc kubenswrapper[4744]: E1008 10:38:14.453868 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:38:25 crc kubenswrapper[4744]: I1008 10:38:25.464522 4744 scope.go:117] "RemoveContainer" containerID="8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685" Oct 08 10:38:25 crc kubenswrapper[4744]: E1008 10:38:25.465208 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:38:38 crc kubenswrapper[4744]: I1008 10:38:38.453693 4744 scope.go:117] "RemoveContainer" containerID="8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685" Oct 08 10:38:38 crc kubenswrapper[4744]: E1008 10:38:38.454424 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:38:53 crc kubenswrapper[4744]: I1008 10:38:53.453818 4744 scope.go:117] "RemoveContainer" containerID="8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685" Oct 08 10:38:53 crc kubenswrapper[4744]: E1008 10:38:53.454740 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:39:00 crc kubenswrapper[4744]: I1008 10:39:00.180849 4744 scope.go:117] "RemoveContainer" containerID="d2468b2f263f6015723219012aab00a7f158f4ad6c0b1a82e526868be5133503" Oct 08 10:39:07 crc kubenswrapper[4744]: I1008 10:39:07.453793 4744 scope.go:117] "RemoveContainer" containerID="8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685" Oct 08 10:39:07 crc kubenswrapper[4744]: E1008 10:39:07.454491 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:39:15 crc kubenswrapper[4744]: I1008 10:39:15.889321 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-thjgq"] Oct 08 10:39:15 crc kubenswrapper[4744]: E1008 10:39:15.891531 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d979fefe-1861-4010-85b7-b709af321101" containerName="extract-utilities" Oct 08 10:39:15 crc kubenswrapper[4744]: I1008 10:39:15.891659 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d979fefe-1861-4010-85b7-b709af321101" containerName="extract-utilities" Oct 08 10:39:15 crc kubenswrapper[4744]: E1008 10:39:15.891754 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d979fefe-1861-4010-85b7-b709af321101" containerName="registry-server" Oct 08 10:39:15 crc kubenswrapper[4744]: I1008 10:39:15.891835 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d979fefe-1861-4010-85b7-b709af321101" containerName="registry-server" Oct 08 10:39:15 crc kubenswrapper[4744]: E1008 10:39:15.891981 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de13d6e4-3524-4157-837a-eec911bc5316" containerName="gather" Oct 08 10:39:15 crc kubenswrapper[4744]: I1008 10:39:15.892069 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="de13d6e4-3524-4157-837a-eec911bc5316" containerName="gather" Oct 08 10:39:15 crc kubenswrapper[4744]: E1008 10:39:15.892158 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d979fefe-1861-4010-85b7-b709af321101" containerName="extract-content" Oct 08 10:39:15 crc kubenswrapper[4744]: I1008 10:39:15.892247 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d979fefe-1861-4010-85b7-b709af321101" containerName="extract-content" Oct 08 10:39:15 crc kubenswrapper[4744]: E1008 10:39:15.892362 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de13d6e4-3524-4157-837a-eec911bc5316" containerName="copy" Oct 08 10:39:15 crc kubenswrapper[4744]: I1008 10:39:15.892489 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="de13d6e4-3524-4157-837a-eec911bc5316" containerName="copy" Oct 08 10:39:15 crc kubenswrapper[4744]: I1008 10:39:15.892837 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="de13d6e4-3524-4157-837a-eec911bc5316" containerName="gather" Oct 08 10:39:15 crc kubenswrapper[4744]: I1008 10:39:15.892956 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d979fefe-1861-4010-85b7-b709af321101" containerName="registry-server" Oct 08 10:39:15 crc kubenswrapper[4744]: I1008 10:39:15.893053 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="de13d6e4-3524-4157-837a-eec911bc5316" containerName="copy" Oct 08 10:39:15 crc kubenswrapper[4744]: I1008 10:39:15.894983 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:15 crc kubenswrapper[4744]: I1008 10:39:15.913803 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-thjgq"] Oct 08 10:39:16 crc kubenswrapper[4744]: I1008 10:39:16.032219 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d46q5\" (UniqueName: \"kubernetes.io/projected/7afa8299-eac1-4944-8472-ce1d53749cbb-kube-api-access-d46q5\") pod \"redhat-marketplace-thjgq\" (UID: \"7afa8299-eac1-4944-8472-ce1d53749cbb\") " pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:16 crc kubenswrapper[4744]: I1008 10:39:16.032352 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7afa8299-eac1-4944-8472-ce1d53749cbb-utilities\") pod \"redhat-marketplace-thjgq\" (UID: \"7afa8299-eac1-4944-8472-ce1d53749cbb\") " pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:16 crc kubenswrapper[4744]: I1008 10:39:16.033035 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7afa8299-eac1-4944-8472-ce1d53749cbb-catalog-content\") pod \"redhat-marketplace-thjgq\" (UID: \"7afa8299-eac1-4944-8472-ce1d53749cbb\") " pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:16 crc kubenswrapper[4744]: I1008 10:39:16.134592 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d46q5\" (UniqueName: \"kubernetes.io/projected/7afa8299-eac1-4944-8472-ce1d53749cbb-kube-api-access-d46q5\") pod \"redhat-marketplace-thjgq\" (UID: \"7afa8299-eac1-4944-8472-ce1d53749cbb\") " pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:16 crc kubenswrapper[4744]: I1008 10:39:16.134647 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7afa8299-eac1-4944-8472-ce1d53749cbb-utilities\") pod \"redhat-marketplace-thjgq\" (UID: \"7afa8299-eac1-4944-8472-ce1d53749cbb\") " pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:16 crc kubenswrapper[4744]: I1008 10:39:16.134686 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7afa8299-eac1-4944-8472-ce1d53749cbb-catalog-content\") pod \"redhat-marketplace-thjgq\" (UID: \"7afa8299-eac1-4944-8472-ce1d53749cbb\") " pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:16 crc kubenswrapper[4744]: I1008 10:39:16.135197 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7afa8299-eac1-4944-8472-ce1d53749cbb-utilities\") pod \"redhat-marketplace-thjgq\" (UID: \"7afa8299-eac1-4944-8472-ce1d53749cbb\") " pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:16 crc kubenswrapper[4744]: I1008 10:39:16.135263 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7afa8299-eac1-4944-8472-ce1d53749cbb-catalog-content\") pod \"redhat-marketplace-thjgq\" (UID: \"7afa8299-eac1-4944-8472-ce1d53749cbb\") " pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:16 crc kubenswrapper[4744]: I1008 10:39:16.163199 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d46q5\" (UniqueName: \"kubernetes.io/projected/7afa8299-eac1-4944-8472-ce1d53749cbb-kube-api-access-d46q5\") pod \"redhat-marketplace-thjgq\" (UID: \"7afa8299-eac1-4944-8472-ce1d53749cbb\") " pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:16 crc kubenswrapper[4744]: I1008 10:39:16.240321 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:16 crc kubenswrapper[4744]: I1008 10:39:16.701993 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-thjgq"] Oct 08 10:39:16 crc kubenswrapper[4744]: I1008 10:39:16.883525 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thjgq" event={"ID":"7afa8299-eac1-4944-8472-ce1d53749cbb","Type":"ContainerStarted","Data":"0a655b17b9e41595a9a6f0def959fc3c0a23260d20a0aaa053ee450cc30d684c"} Oct 08 10:39:17 crc kubenswrapper[4744]: I1008 10:39:17.894555 4744 generic.go:334] "Generic (PLEG): container finished" podID="7afa8299-eac1-4944-8472-ce1d53749cbb" containerID="190edcb9ec28bd27c350ad3ace9813f05e8150ae3312acc608428c0a2bc8fe10" exitCode=0 Oct 08 10:39:17 crc kubenswrapper[4744]: I1008 10:39:17.894606 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thjgq" event={"ID":"7afa8299-eac1-4944-8472-ce1d53749cbb","Type":"ContainerDied","Data":"190edcb9ec28bd27c350ad3ace9813f05e8150ae3312acc608428c0a2bc8fe10"} Oct 08 10:39:18 crc kubenswrapper[4744]: I1008 10:39:18.938051 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thjgq" event={"ID":"7afa8299-eac1-4944-8472-ce1d53749cbb","Type":"ContainerStarted","Data":"90a2e91abc4a91b7cb93ab2d9c0f82c9c94827460b2863dfcd5da10084979679"} Oct 08 10:39:19 crc kubenswrapper[4744]: I1008 10:39:19.948640 4744 generic.go:334] "Generic (PLEG): container finished" podID="7afa8299-eac1-4944-8472-ce1d53749cbb" containerID="90a2e91abc4a91b7cb93ab2d9c0f82c9c94827460b2863dfcd5da10084979679" exitCode=0 Oct 08 10:39:19 crc kubenswrapper[4744]: I1008 10:39:19.948824 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thjgq" event={"ID":"7afa8299-eac1-4944-8472-ce1d53749cbb","Type":"ContainerDied","Data":"90a2e91abc4a91b7cb93ab2d9c0f82c9c94827460b2863dfcd5da10084979679"} Oct 08 10:39:20 crc kubenswrapper[4744]: I1008 10:39:20.453065 4744 scope.go:117] "RemoveContainer" containerID="8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685" Oct 08 10:39:20 crc kubenswrapper[4744]: E1008 10:39:20.453669 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" Oct 08 10:39:20 crc kubenswrapper[4744]: I1008 10:39:20.960185 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thjgq" event={"ID":"7afa8299-eac1-4944-8472-ce1d53749cbb","Type":"ContainerStarted","Data":"f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8"} Oct 08 10:39:20 crc kubenswrapper[4744]: I1008 10:39:20.981427 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-thjgq" podStartSLOduration=3.182387103 podStartE2EDuration="5.981412338s" podCreationTimestamp="2025-10-08 10:39:15 +0000 UTC" firstStartedPulling="2025-10-08 10:39:17.897195882 +0000 UTC m=+5253.144841121" lastFinishedPulling="2025-10-08 10:39:20.696221117 +0000 UTC m=+5255.943866356" observedRunningTime="2025-10-08 10:39:20.979803864 +0000 UTC m=+5256.227449113" watchObservedRunningTime="2025-10-08 10:39:20.981412338 +0000 UTC m=+5256.229057577" Oct 08 10:39:26 crc kubenswrapper[4744]: I1008 10:39:26.242225 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:26 crc kubenswrapper[4744]: I1008 10:39:26.242877 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:26 crc kubenswrapper[4744]: I1008 10:39:26.302931 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:27 crc kubenswrapper[4744]: I1008 10:39:27.060124 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:27 crc kubenswrapper[4744]: I1008 10:39:27.107929 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-thjgq"] Oct 08 10:39:29 crc kubenswrapper[4744]: I1008 10:39:29.038687 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-thjgq" podUID="7afa8299-eac1-4944-8472-ce1d53749cbb" containerName="registry-server" containerID="cri-o://f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8" gracePeriod=2 Oct 08 10:39:29 crc kubenswrapper[4744]: I1008 10:39:29.859754 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:29 crc kubenswrapper[4744]: I1008 10:39:29.994337 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7afa8299-eac1-4944-8472-ce1d53749cbb-utilities\") pod \"7afa8299-eac1-4944-8472-ce1d53749cbb\" (UID: \"7afa8299-eac1-4944-8472-ce1d53749cbb\") " Oct 08 10:39:29 crc kubenswrapper[4744]: I1008 10:39:29.994913 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d46q5\" (UniqueName: \"kubernetes.io/projected/7afa8299-eac1-4944-8472-ce1d53749cbb-kube-api-access-d46q5\") pod \"7afa8299-eac1-4944-8472-ce1d53749cbb\" (UID: \"7afa8299-eac1-4944-8472-ce1d53749cbb\") " Oct 08 10:39:29 crc kubenswrapper[4744]: I1008 10:39:29.995061 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7afa8299-eac1-4944-8472-ce1d53749cbb-catalog-content\") pod \"7afa8299-eac1-4944-8472-ce1d53749cbb\" (UID: \"7afa8299-eac1-4944-8472-ce1d53749cbb\") " Oct 08 10:39:29 crc kubenswrapper[4744]: I1008 10:39:29.995443 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7afa8299-eac1-4944-8472-ce1d53749cbb-utilities" (OuterVolumeSpecName: "utilities") pod "7afa8299-eac1-4944-8472-ce1d53749cbb" (UID: "7afa8299-eac1-4944-8472-ce1d53749cbb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:39:29 crc kubenswrapper[4744]: I1008 10:39:29.995963 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7afa8299-eac1-4944-8472-ce1d53749cbb-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.001962 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7afa8299-eac1-4944-8472-ce1d53749cbb-kube-api-access-d46q5" (OuterVolumeSpecName: "kube-api-access-d46q5") pod "7afa8299-eac1-4944-8472-ce1d53749cbb" (UID: "7afa8299-eac1-4944-8472-ce1d53749cbb"). InnerVolumeSpecName "kube-api-access-d46q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.008620 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7afa8299-eac1-4944-8472-ce1d53749cbb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7afa8299-eac1-4944-8472-ce1d53749cbb" (UID: "7afa8299-eac1-4944-8472-ce1d53749cbb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.056759 4744 generic.go:334] "Generic (PLEG): container finished" podID="7afa8299-eac1-4944-8472-ce1d53749cbb" containerID="f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8" exitCode=0 Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.056863 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thjgq" event={"ID":"7afa8299-eac1-4944-8472-ce1d53749cbb","Type":"ContainerDied","Data":"f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8"} Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.056899 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-thjgq" event={"ID":"7afa8299-eac1-4944-8472-ce1d53749cbb","Type":"ContainerDied","Data":"0a655b17b9e41595a9a6f0def959fc3c0a23260d20a0aaa053ee450cc30d684c"} Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.056917 4744 scope.go:117] "RemoveContainer" containerID="f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.057518 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-thjgq" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.087767 4744 scope.go:117] "RemoveContainer" containerID="90a2e91abc4a91b7cb93ab2d9c0f82c9c94827460b2863dfcd5da10084979679" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.093561 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-thjgq"] Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.103637 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-thjgq"] Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.105406 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d46q5\" (UniqueName: \"kubernetes.io/projected/7afa8299-eac1-4944-8472-ce1d53749cbb-kube-api-access-d46q5\") on node \"crc\" DevicePath \"\"" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.105430 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7afa8299-eac1-4944-8472-ce1d53749cbb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.109244 4744 scope.go:117] "RemoveContainer" containerID="190edcb9ec28bd27c350ad3ace9813f05e8150ae3312acc608428c0a2bc8fe10" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.156676 4744 scope.go:117] "RemoveContainer" containerID="f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8" Oct 08 10:39:30 crc kubenswrapper[4744]: E1008 10:39:30.157607 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8\": container with ID starting with f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8 not found: ID does not exist" containerID="f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.157640 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8"} err="failed to get container status \"f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8\": rpc error: code = NotFound desc = could not find container \"f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8\": container with ID starting with f5a21355610d3eae524517e8563e3fcac7a2677ff425db95b9406ef5ae0748e8 not found: ID does not exist" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.157666 4744 scope.go:117] "RemoveContainer" containerID="90a2e91abc4a91b7cb93ab2d9c0f82c9c94827460b2863dfcd5da10084979679" Oct 08 10:39:30 crc kubenswrapper[4744]: E1008 10:39:30.158004 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90a2e91abc4a91b7cb93ab2d9c0f82c9c94827460b2863dfcd5da10084979679\": container with ID starting with 90a2e91abc4a91b7cb93ab2d9c0f82c9c94827460b2863dfcd5da10084979679 not found: ID does not exist" containerID="90a2e91abc4a91b7cb93ab2d9c0f82c9c94827460b2863dfcd5da10084979679" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.158024 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90a2e91abc4a91b7cb93ab2d9c0f82c9c94827460b2863dfcd5da10084979679"} err="failed to get container status \"90a2e91abc4a91b7cb93ab2d9c0f82c9c94827460b2863dfcd5da10084979679\": rpc error: code = NotFound desc = could not find container \"90a2e91abc4a91b7cb93ab2d9c0f82c9c94827460b2863dfcd5da10084979679\": container with ID starting with 90a2e91abc4a91b7cb93ab2d9c0f82c9c94827460b2863dfcd5da10084979679 not found: ID does not exist" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.158039 4744 scope.go:117] "RemoveContainer" containerID="190edcb9ec28bd27c350ad3ace9813f05e8150ae3312acc608428c0a2bc8fe10" Oct 08 10:39:30 crc kubenswrapper[4744]: E1008 10:39:30.158913 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"190edcb9ec28bd27c350ad3ace9813f05e8150ae3312acc608428c0a2bc8fe10\": container with ID starting with 190edcb9ec28bd27c350ad3ace9813f05e8150ae3312acc608428c0a2bc8fe10 not found: ID does not exist" containerID="190edcb9ec28bd27c350ad3ace9813f05e8150ae3312acc608428c0a2bc8fe10" Oct 08 10:39:30 crc kubenswrapper[4744]: I1008 10:39:30.158959 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"190edcb9ec28bd27c350ad3ace9813f05e8150ae3312acc608428c0a2bc8fe10"} err="failed to get container status \"190edcb9ec28bd27c350ad3ace9813f05e8150ae3312acc608428c0a2bc8fe10\": rpc error: code = NotFound desc = could not find container \"190edcb9ec28bd27c350ad3ace9813f05e8150ae3312acc608428c0a2bc8fe10\": container with ID starting with 190edcb9ec28bd27c350ad3ace9813f05e8150ae3312acc608428c0a2bc8fe10 not found: ID does not exist" Oct 08 10:39:31 crc kubenswrapper[4744]: I1008 10:39:31.470494 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7afa8299-eac1-4944-8472-ce1d53749cbb" path="/var/lib/kubelet/pods/7afa8299-eac1-4944-8472-ce1d53749cbb/volumes" Oct 08 10:39:32 crc kubenswrapper[4744]: I1008 10:39:32.453005 4744 scope.go:117] "RemoveContainer" containerID="8057eea7316cd724daf49034ed4da5af886bd2929f736b4e55e115ddcdecb685" Oct 08 10:39:32 crc kubenswrapper[4744]: E1008 10:39:32.453320 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ttxzq_openshift-machine-config-operator(9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb)\"" pod="openshift-machine-config-operator/machine-config-daemon-ttxzq" podUID="9d1b4cf3-79bd-4fc9-8718-60c1ea87fadb" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515071437555024461 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015071437556017377 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015071424674016517 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015071424674015467 5ustar corecore